00:00:00.000 Started by upstream project "autotest-per-patch" build number 132416 00:00:00.000 originally caused by: 00:00:00.000 Started by user sys_sgci 00:00:00.015 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-vg.groovy 00:00:00.015 The recommended git tool is: git 00:00:00.016 using credential 00000000-0000-0000-0000-000000000002 00:00:00.017 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:00.028 Fetching changes from the remote Git repository 00:00:00.032 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:00.048 Using shallow fetch with depth 1 00:00:00.048 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:00.048 > git --version # timeout=10 00:00:00.064 > git --version # 'git version 2.39.2' 00:00:00.064 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:00.097 Setting http proxy: proxy-dmz.intel.com:911 00:00:00.097 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:03.042 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:03.055 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:03.070 Checking out Revision db4637e8b949f278f369ec13f70585206ccd9507 (FETCH_HEAD) 00:00:03.070 > git config core.sparsecheckout # timeout=10 00:00:03.085 > git read-tree -mu HEAD # timeout=10 00:00:03.103 > git checkout -f db4637e8b949f278f369ec13f70585206ccd9507 # timeout=5 00:00:03.129 Commit message: "jenkins/jjb-config: Add missing SPDK_TEST_NVME_INTERRUPT flag" 00:00:03.129 > git rev-list --no-walk db4637e8b949f278f369ec13f70585206ccd9507 # timeout=10 00:00:03.227 [Pipeline] Start of Pipeline 00:00:03.242 [Pipeline] library 00:00:03.244 Loading library shm_lib@master 00:00:07.671 Library shm_lib@master is cached. Copying from home. 00:00:07.758 [Pipeline] node 00:00:22.800 Still waiting to schedule task 00:00:22.800 Waiting for next available executor on ‘vagrant-vm-host’ 00:03:29.379 Running on VM-host-SM38 in /var/jenkins/workspace/raid-vg-autotest 00:03:29.381 [Pipeline] { 00:03:29.391 [Pipeline] catchError 00:03:29.393 [Pipeline] { 00:03:29.408 [Pipeline] wrap 00:03:29.418 [Pipeline] { 00:03:29.423 [Pipeline] stage 00:03:29.425 [Pipeline] { (Prologue) 00:03:29.439 [Pipeline] echo 00:03:29.440 Node: VM-host-SM38 00:03:29.445 [Pipeline] cleanWs 00:03:29.465 [WS-CLEANUP] Deleting project workspace... 00:03:29.465 [WS-CLEANUP] Deferred wipeout is used... 00:03:29.472 [WS-CLEANUP] done 00:03:29.700 [Pipeline] setCustomBuildProperty 00:03:29.797 [Pipeline] httpRequest 00:03:30.201 [Pipeline] echo 00:03:30.203 Sorcerer 10.211.164.20 is alive 00:03:30.215 [Pipeline] retry 00:03:30.217 [Pipeline] { 00:03:30.232 [Pipeline] httpRequest 00:03:30.237 HttpMethod: GET 00:03:30.238 URL: http://10.211.164.20/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:03:30.239 Sending request to url: http://10.211.164.20/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:03:30.239 Response Code: HTTP/1.1 200 OK 00:03:30.240 Success: Status code 200 is in the accepted range: 200,404 00:03:30.240 Saving response body to /var/jenkins/workspace/raid-vg-autotest/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:03:30.901 [Pipeline] } 00:03:30.918 [Pipeline] // retry 00:03:30.926 [Pipeline] sh 00:03:31.203 + tar --no-same-owner -xf jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:03:31.219 [Pipeline] httpRequest 00:03:31.587 [Pipeline] echo 00:03:31.589 Sorcerer 10.211.164.20 is alive 00:03:31.597 [Pipeline] retry 00:03:31.599 [Pipeline] { 00:03:31.614 [Pipeline] httpRequest 00:03:31.618 HttpMethod: GET 00:03:31.619 URL: http://10.211.164.20/packages/spdk_0728de5b0db32c537468e1c1f0bb2b85c9971877.tar.gz 00:03:31.620 Sending request to url: http://10.211.164.20/packages/spdk_0728de5b0db32c537468e1c1f0bb2b85c9971877.tar.gz 00:03:31.620 Response Code: HTTP/1.1 200 OK 00:03:31.621 Success: Status code 200 is in the accepted range: 200,404 00:03:31.622 Saving response body to /var/jenkins/workspace/raid-vg-autotest/spdk_0728de5b0db32c537468e1c1f0bb2b85c9971877.tar.gz 00:03:38.662 [Pipeline] } 00:03:38.680 [Pipeline] // retry 00:03:38.688 [Pipeline] sh 00:03:38.993 + tar --no-same-owner -xf spdk_0728de5b0db32c537468e1c1f0bb2b85c9971877.tar.gz 00:03:41.546 [Pipeline] sh 00:03:41.822 + git -C spdk log --oneline -n5 00:03:41.822 0728de5b0 nvmf: Add hide_metadata option to nvmf_subsystem_add_ns 00:03:41.822 349af566b nvmf: Get metadata config by not bdev but bdev_desc 00:03:41.822 1981e6eec bdevperf: Add hide_metadata option 00:03:41.822 66a383faf bdevperf: Get metadata config by not bdev but bdev_desc 00:03:41.822 25916e30c bdevperf: Store the result of DIF type check into job structure 00:03:41.839 [Pipeline] writeFile 00:03:41.852 [Pipeline] sh 00:03:42.131 + jbp/jenkins/jjb-config/jobs/scripts/autorun_quirks.sh 00:03:42.141 [Pipeline] sh 00:03:42.420 + cat autorun-spdk.conf 00:03:42.421 SPDK_RUN_FUNCTIONAL_TEST=1 00:03:42.421 SPDK_RUN_ASAN=1 00:03:42.421 SPDK_RUN_UBSAN=1 00:03:42.421 SPDK_TEST_RAID=1 00:03:42.421 SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:03:42.426 RUN_NIGHTLY=0 00:03:42.429 [Pipeline] } 00:03:42.443 [Pipeline] // stage 00:03:42.459 [Pipeline] stage 00:03:42.461 [Pipeline] { (Run VM) 00:03:42.475 [Pipeline] sh 00:03:42.753 + jbp/jenkins/jjb-config/jobs/scripts/prepare_nvme.sh 00:03:42.753 + echo 'Start stage prepare_nvme.sh' 00:03:42.753 Start stage prepare_nvme.sh 00:03:42.753 + [[ -n 2 ]] 00:03:42.753 + disk_prefix=ex2 00:03:42.753 + [[ -n /var/jenkins/workspace/raid-vg-autotest ]] 00:03:42.753 + [[ -e /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf ]] 00:03:42.753 + source /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf 00:03:42.753 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:03:42.753 ++ SPDK_RUN_ASAN=1 00:03:42.753 ++ SPDK_RUN_UBSAN=1 00:03:42.753 ++ SPDK_TEST_RAID=1 00:03:42.753 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:03:42.753 ++ RUN_NIGHTLY=0 00:03:42.753 + cd /var/jenkins/workspace/raid-vg-autotest 00:03:42.753 + nvme_files=() 00:03:42.753 + declare -A nvme_files 00:03:42.753 + backend_dir=/var/lib/libvirt/images/backends 00:03:42.753 + nvme_files['nvme.img']=5G 00:03:42.753 + nvme_files['nvme-cmb.img']=5G 00:03:42.753 + nvme_files['nvme-multi0.img']=4G 00:03:42.753 + nvme_files['nvme-multi1.img']=4G 00:03:42.753 + nvme_files['nvme-multi2.img']=4G 00:03:42.753 + nvme_files['nvme-openstack.img']=8G 00:03:42.753 + nvme_files['nvme-zns.img']=5G 00:03:42.753 + (( SPDK_TEST_NVME_PMR == 1 )) 00:03:42.753 + (( SPDK_TEST_FTL == 1 )) 00:03:42.753 + (( SPDK_TEST_NVME_FDP == 1 )) 00:03:42.753 + [[ ! -d /var/lib/libvirt/images/backends ]] 00:03:42.753 + for nvme in "${!nvme_files[@]}" 00:03:42.753 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex2-nvme-multi2.img -s 4G 00:03:42.753 Formatting '/var/lib/libvirt/images/backends/ex2-nvme-multi2.img', fmt=raw size=4294967296 preallocation=falloc 00:03:42.753 + for nvme in "${!nvme_files[@]}" 00:03:42.753 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex2-nvme-cmb.img -s 5G 00:03:42.753 Formatting '/var/lib/libvirt/images/backends/ex2-nvme-cmb.img', fmt=raw size=5368709120 preallocation=falloc 00:03:42.753 + for nvme in "${!nvme_files[@]}" 00:03:42.753 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex2-nvme-openstack.img -s 8G 00:03:42.753 Formatting '/var/lib/libvirt/images/backends/ex2-nvme-openstack.img', fmt=raw size=8589934592 preallocation=falloc 00:03:42.753 + for nvme in "${!nvme_files[@]}" 00:03:42.753 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex2-nvme-zns.img -s 5G 00:03:42.753 Formatting '/var/lib/libvirt/images/backends/ex2-nvme-zns.img', fmt=raw size=5368709120 preallocation=falloc 00:03:42.753 + for nvme in "${!nvme_files[@]}" 00:03:42.753 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex2-nvme-multi1.img -s 4G 00:03:42.753 Formatting '/var/lib/libvirt/images/backends/ex2-nvme-multi1.img', fmt=raw size=4294967296 preallocation=falloc 00:03:42.753 + for nvme in "${!nvme_files[@]}" 00:03:42.753 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex2-nvme-multi0.img -s 4G 00:03:43.011 Formatting '/var/lib/libvirt/images/backends/ex2-nvme-multi0.img', fmt=raw size=4294967296 preallocation=falloc 00:03:43.011 + for nvme in "${!nvme_files[@]}" 00:03:43.011 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex2-nvme.img -s 5G 00:03:43.011 Formatting '/var/lib/libvirt/images/backends/ex2-nvme.img', fmt=raw size=5368709120 preallocation=falloc 00:03:43.011 ++ sudo grep -rl ex2-nvme.img /etc/libvirt/qemu 00:03:43.011 + echo 'End stage prepare_nvme.sh' 00:03:43.011 End stage prepare_nvme.sh 00:03:43.022 [Pipeline] sh 00:03:43.301 + DISTRO=fedora39 00:03:43.301 + CPUS=10 00:03:43.301 + RAM=12288 00:03:43.301 + jbp/jenkins/jjb-config/jobs/scripts/vagrant_create_vm.sh 00:03:43.301 Setup: -n 10 -s 12288 -x -p libvirt --qemu-emulator=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 --nic-model=e1000 -b /var/lib/libvirt/images/backends/ex2-nvme.img -b /var/lib/libvirt/images/backends/ex2-nvme-multi0.img,nvme,/var/lib/libvirt/images/backends/ex2-nvme-multi1.img:/var/lib/libvirt/images/backends/ex2-nvme-multi2.img -H -a -v -f fedora39 00:03:43.301 00:03:43.301 DIR=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant 00:03:43.301 SPDK_DIR=/var/jenkins/workspace/raid-vg-autotest/spdk 00:03:43.301 VAGRANT_TARGET=/var/jenkins/workspace/raid-vg-autotest 00:03:43.301 HELP=0 00:03:43.301 DRY_RUN=0 00:03:43.301 NVME_FILE=/var/lib/libvirt/images/backends/ex2-nvme.img,/var/lib/libvirt/images/backends/ex2-nvme-multi0.img, 00:03:43.301 NVME_DISKS_TYPE=nvme,nvme, 00:03:43.301 NVME_AUTO_CREATE=0 00:03:43.301 NVME_DISKS_NAMESPACES=,/var/lib/libvirt/images/backends/ex2-nvme-multi1.img:/var/lib/libvirt/images/backends/ex2-nvme-multi2.img, 00:03:43.301 NVME_CMB=,, 00:03:43.301 NVME_PMR=,, 00:03:43.301 NVME_ZNS=,, 00:03:43.301 NVME_MS=,, 00:03:43.301 NVME_FDP=,, 00:03:43.301 SPDK_VAGRANT_DISTRO=fedora39 00:03:43.301 SPDK_VAGRANT_VMCPU=10 00:03:43.301 SPDK_VAGRANT_VMRAM=12288 00:03:43.301 SPDK_VAGRANT_PROVIDER=libvirt 00:03:43.301 SPDK_VAGRANT_HTTP_PROXY= 00:03:43.301 SPDK_QEMU_EMULATOR=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 00:03:43.301 SPDK_OPENSTACK_NETWORK=0 00:03:43.301 VAGRANT_PACKAGE_BOX=0 00:03:43.301 VAGRANTFILE=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant/Vagrantfile 00:03:43.301 FORCE_DISTRO=true 00:03:43.301 VAGRANT_BOX_VERSION= 00:03:43.301 EXTRA_VAGRANTFILES= 00:03:43.301 NIC_MODEL=e1000 00:03:43.301 00:03:43.301 mkdir: created directory '/var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt' 00:03:43.301 /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt /var/jenkins/workspace/raid-vg-autotest 00:03:45.825 Bringing machine 'default' up with 'libvirt' provider... 00:03:46.083 ==> default: Creating image (snapshot of base box volume). 00:03:46.341 ==> default: Creating domain with the following settings... 00:03:46.341 ==> default: -- Name: fedora39-39-1.5-1721788873-2326_default_1732117904_235040679d95c83ab5fa 00:03:46.341 ==> default: -- Domain type: kvm 00:03:46.341 ==> default: -- Cpus: 10 00:03:46.341 ==> default: -- Feature: acpi 00:03:46.341 ==> default: -- Feature: apic 00:03:46.341 ==> default: -- Feature: pae 00:03:46.341 ==> default: -- Memory: 12288M 00:03:46.341 ==> default: -- Memory Backing: hugepages: 00:03:46.341 ==> default: -- Management MAC: 00:03:46.341 ==> default: -- Loader: 00:03:46.341 ==> default: -- Nvram: 00:03:46.341 ==> default: -- Base box: spdk/fedora39 00:03:46.341 ==> default: -- Storage pool: default 00:03:46.341 ==> default: -- Image: /var/lib/libvirt/images/fedora39-39-1.5-1721788873-2326_default_1732117904_235040679d95c83ab5fa.img (20G) 00:03:46.341 ==> default: -- Volume Cache: default 00:03:46.341 ==> default: -- Kernel: 00:03:46.341 ==> default: -- Initrd: 00:03:46.341 ==> default: -- Graphics Type: vnc 00:03:46.341 ==> default: -- Graphics Port: -1 00:03:46.341 ==> default: -- Graphics IP: 127.0.0.1 00:03:46.341 ==> default: -- Graphics Password: Not defined 00:03:46.341 ==> default: -- Video Type: cirrus 00:03:46.341 ==> default: -- Video VRAM: 9216 00:03:46.341 ==> default: -- Sound Type: 00:03:46.341 ==> default: -- Keymap: en-us 00:03:46.341 ==> default: -- TPM Path: 00:03:46.341 ==> default: -- INPUT: type=mouse, bus=ps2 00:03:46.341 ==> default: -- Command line args: 00:03:46.341 ==> default: -> value=-device, 00:03:46.341 ==> default: -> value=nvme,id=nvme-0,serial=12340,addr=0x10, 00:03:46.341 ==> default: -> value=-drive, 00:03:46.342 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex2-nvme.img,if=none,id=nvme-0-drive0, 00:03:46.342 ==> default: -> value=-device, 00:03:46.342 ==> default: -> value=nvme-ns,drive=nvme-0-drive0,bus=nvme-0,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:03:46.342 ==> default: -> value=-device, 00:03:46.342 ==> default: -> value=nvme,id=nvme-1,serial=12341,addr=0x11, 00:03:46.342 ==> default: -> value=-drive, 00:03:46.342 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex2-nvme-multi0.img,if=none,id=nvme-1-drive0, 00:03:46.342 ==> default: -> value=-device, 00:03:46.342 ==> default: -> value=nvme-ns,drive=nvme-1-drive0,bus=nvme-1,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:03:46.342 ==> default: -> value=-drive, 00:03:46.342 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex2-nvme-multi1.img,if=none,id=nvme-1-drive1, 00:03:46.342 ==> default: -> value=-device, 00:03:46.342 ==> default: -> value=nvme-ns,drive=nvme-1-drive1,bus=nvme-1,nsid=2,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:03:46.342 ==> default: -> value=-drive, 00:03:46.342 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex2-nvme-multi2.img,if=none,id=nvme-1-drive2, 00:03:46.342 ==> default: -> value=-device, 00:03:46.342 ==> default: -> value=nvme-ns,drive=nvme-1-drive2,bus=nvme-1,nsid=3,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:03:46.342 ==> default: Creating shared folders metadata... 00:03:46.600 ==> default: Starting domain. 00:03:47.972 ==> default: Waiting for domain to get an IP address... 00:04:02.954 ==> default: Waiting for SSH to become available... 00:04:02.954 ==> default: Configuring and enabling network interfaces... 00:04:06.234 default: SSH address: 192.168.121.43:22 00:04:06.234 default: SSH username: vagrant 00:04:06.234 default: SSH auth method: private key 00:04:07.605 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/spdk/ => /home/vagrant/spdk_repo/spdk 00:04:14.267 ==> default: Mounting SSHFS shared folder... 00:04:15.199 ==> default: Mounting folder via SSHFS: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/fedora39-libvirt/output => /home/vagrant/spdk_repo/output 00:04:15.199 ==> default: Checking Mount.. 00:04:16.579 ==> default: Folder Successfully Mounted! 00:04:16.579 00:04:16.579 SUCCESS! 00:04:16.579 00:04:16.579 cd to /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt and type "vagrant ssh" to use. 00:04:16.579 Use vagrant "suspend" and vagrant "resume" to stop and start. 00:04:16.579 Use vagrant "destroy" followed by "rm -rf /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt" to destroy all trace of vm. 00:04:16.579 00:04:16.587 [Pipeline] } 00:04:16.604 [Pipeline] // stage 00:04:16.614 [Pipeline] dir 00:04:16.615 Running in /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt 00:04:16.616 [Pipeline] { 00:04:16.629 [Pipeline] catchError 00:04:16.631 [Pipeline] { 00:04:16.646 [Pipeline] sh 00:04:16.927 + vagrant ssh-config --host vagrant 00:04:16.927 + tee ssh_conf 00:04:16.927 + sed -ne '/^Host/,$p' 00:04:19.451 Host vagrant 00:04:19.451 HostName 192.168.121.43 00:04:19.451 User vagrant 00:04:19.451 Port 22 00:04:19.451 UserKnownHostsFile /dev/null 00:04:19.451 StrictHostKeyChecking no 00:04:19.451 PasswordAuthentication no 00:04:19.451 IdentityFile /var/lib/libvirt/images/.vagrant.d/boxes/spdk-VAGRANTSLASH-fedora39/39-1.5-1721788873-2326/libvirt/fedora39 00:04:19.451 IdentitiesOnly yes 00:04:19.451 LogLevel FATAL 00:04:19.451 ForwardAgent yes 00:04:19.451 ForwardX11 yes 00:04:19.451 00:04:19.462 [Pipeline] withEnv 00:04:19.463 [Pipeline] { 00:04:19.473 [Pipeline] sh 00:04:19.745 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant '#!/bin/bash 00:04:19.745 source /etc/os-release 00:04:19.745 [[ -e /image.version ]] && img=$(< /image.version) 00:04:19.745 # Minimal, systemd-like check. 00:04:19.745 if [[ -e /.dockerenv ]]; then 00:04:19.745 # Clear garbage from the node'\''s name: 00:04:19.745 # agt-er_autotest_547-896 -> autotest_547-896 00:04:19.745 # $HOSTNAME is the actual container id 00:04:19.745 agent=$HOSTNAME@${DOCKER_SWARM_PLUGIN_JENKINS_AGENT_NAME#*_} 00:04:19.745 if grep -q "/etc/hostname" /proc/self/mountinfo; then 00:04:19.745 # We can assume this is a mount from a host where container is running, 00:04:19.745 # so fetch its hostname to easily identify the target swarm worker. 00:04:19.745 container="$(< /etc/hostname) ($agent)" 00:04:19.745 else 00:04:19.745 # Fallback 00:04:19.745 container=$agent 00:04:19.745 fi 00:04:19.745 fi 00:04:19.745 echo "${NAME} ${VERSION_ID}|$(uname -r)|${img:-N/A}|${container:-N/A}" 00:04:19.745 ' 00:04:19.754 [Pipeline] } 00:04:19.771 [Pipeline] // withEnv 00:04:19.778 [Pipeline] setCustomBuildProperty 00:04:19.793 [Pipeline] stage 00:04:19.795 [Pipeline] { (Tests) 00:04:19.811 [Pipeline] sh 00:04:20.084 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh vagrant@vagrant:./ 00:04:20.096 [Pipeline] sh 00:04:20.487 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/pkgdep-autoruner.sh vagrant@vagrant:./ 00:04:20.501 [Pipeline] timeout 00:04:20.501 Timeout set to expire in 1 hr 30 min 00:04:20.503 [Pipeline] { 00:04:20.518 [Pipeline] sh 00:04:20.793 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant 'git -C spdk_repo/spdk reset --hard' 00:04:21.051 HEAD is now at 0728de5b0 nvmf: Add hide_metadata option to nvmf_subsystem_add_ns 00:04:21.063 [Pipeline] sh 00:04:21.344 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant 'sudo chown vagrant:vagrant spdk_repo' 00:04:21.359 [Pipeline] sh 00:04:21.636 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf vagrant@vagrant:spdk_repo 00:04:21.653 [Pipeline] sh 00:04:21.926 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant 'JOB_BASE_NAME=raid-vg-autotest ./autoruner.sh spdk_repo' 00:04:21.926 ++ readlink -f spdk_repo 00:04:21.926 + DIR_ROOT=/home/vagrant/spdk_repo 00:04:21.926 + [[ -n /home/vagrant/spdk_repo ]] 00:04:21.926 + DIR_SPDK=/home/vagrant/spdk_repo/spdk 00:04:21.926 + DIR_OUTPUT=/home/vagrant/spdk_repo/output 00:04:21.926 + [[ -d /home/vagrant/spdk_repo/spdk ]] 00:04:21.926 + [[ ! -d /home/vagrant/spdk_repo/output ]] 00:04:21.926 + [[ -d /home/vagrant/spdk_repo/output ]] 00:04:21.926 + [[ raid-vg-autotest == pkgdep-* ]] 00:04:21.926 + cd /home/vagrant/spdk_repo 00:04:21.926 + source /etc/os-release 00:04:21.926 ++ NAME='Fedora Linux' 00:04:21.926 ++ VERSION='39 (Cloud Edition)' 00:04:21.926 ++ ID=fedora 00:04:21.926 ++ VERSION_ID=39 00:04:21.926 ++ VERSION_CODENAME= 00:04:21.926 ++ PLATFORM_ID=platform:f39 00:04:21.926 ++ PRETTY_NAME='Fedora Linux 39 (Cloud Edition)' 00:04:21.926 ++ ANSI_COLOR='0;38;2;60;110;180' 00:04:21.926 ++ LOGO=fedora-logo-icon 00:04:21.926 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:39 00:04:21.926 ++ HOME_URL=https://fedoraproject.org/ 00:04:21.926 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f39/system-administrators-guide/ 00:04:21.926 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:04:21.926 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:04:21.926 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:04:21.926 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=39 00:04:21.926 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:04:21.926 ++ REDHAT_SUPPORT_PRODUCT_VERSION=39 00:04:21.927 ++ SUPPORT_END=2024-11-12 00:04:21.927 ++ VARIANT='Cloud Edition' 00:04:21.927 ++ VARIANT_ID=cloud 00:04:21.927 + uname -a 00:04:21.927 Linux fedora39-cloud-1721788873-2326 6.8.9-200.fc39.x86_64 #1 SMP PREEMPT_DYNAMIC Wed Jul 24 03:04:40 UTC 2024 x86_64 GNU/Linux 00:04:21.927 + sudo /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:04:22.495 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:04:22.496 Hugepages 00:04:22.496 node hugesize free / total 00:04:22.496 node0 1048576kB 0 / 0 00:04:22.496 node0 2048kB 0 / 0 00:04:22.496 00:04:22.496 Type BDF Vendor Device NUMA Driver Device Block devices 00:04:22.496 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:04:22.496 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:04:22.496 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:04:22.496 + rm -f /tmp/spdk-ld-path 00:04:22.496 + source autorun-spdk.conf 00:04:22.496 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:04:22.496 ++ SPDK_RUN_ASAN=1 00:04:22.496 ++ SPDK_RUN_UBSAN=1 00:04:22.496 ++ SPDK_TEST_RAID=1 00:04:22.496 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:04:22.496 ++ RUN_NIGHTLY=0 00:04:22.496 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:04:22.496 + [[ -n '' ]] 00:04:22.496 + sudo git config --global --add safe.directory /home/vagrant/spdk_repo/spdk 00:04:22.496 + for M in /var/spdk/build-*-manifest.txt 00:04:22.496 + [[ -f /var/spdk/build-kernel-manifest.txt ]] 00:04:22.496 + cp /var/spdk/build-kernel-manifest.txt /home/vagrant/spdk_repo/output/ 00:04:22.496 + for M in /var/spdk/build-*-manifest.txt 00:04:22.496 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:04:22.496 + cp /var/spdk/build-pkg-manifest.txt /home/vagrant/spdk_repo/output/ 00:04:22.496 + for M in /var/spdk/build-*-manifest.txt 00:04:22.496 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:04:22.496 + cp /var/spdk/build-repo-manifest.txt /home/vagrant/spdk_repo/output/ 00:04:22.496 ++ uname 00:04:22.496 + [[ Linux == \L\i\n\u\x ]] 00:04:22.496 + sudo dmesg -T 00:04:22.801 + sudo dmesg --clear 00:04:22.801 + dmesg_pid=4975 00:04:22.801 + [[ Fedora Linux == FreeBSD ]] 00:04:22.801 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:04:22.801 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:04:22.801 + sudo dmesg -Tw 00:04:22.801 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:04:22.801 + [[ -x /usr/src/fio-static/fio ]] 00:04:22.801 + export FIO_BIN=/usr/src/fio-static/fio 00:04:22.801 + FIO_BIN=/usr/src/fio-static/fio 00:04:22.801 + [[ '' == \/\q\e\m\u\_\v\f\i\o\/* ]] 00:04:22.801 + [[ ! -v VFIO_QEMU_BIN ]] 00:04:22.801 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:04:22.801 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:04:22.801 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:04:22.801 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:04:22.801 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:04:22.801 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:04:22.801 + spdk/autorun.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:04:22.801 15:52:20 -- common/autotest_common.sh@1692 -- $ [[ n == y ]] 00:04:22.801 15:52:20 -- spdk/autorun.sh@20 -- $ source /home/vagrant/spdk_repo/autorun-spdk.conf 00:04:22.801 15:52:20 -- spdk_repo/autorun-spdk.conf@1 -- $ SPDK_RUN_FUNCTIONAL_TEST=1 00:04:22.801 15:52:20 -- spdk_repo/autorun-spdk.conf@2 -- $ SPDK_RUN_ASAN=1 00:04:22.801 15:52:20 -- spdk_repo/autorun-spdk.conf@3 -- $ SPDK_RUN_UBSAN=1 00:04:22.801 15:52:20 -- spdk_repo/autorun-spdk.conf@4 -- $ SPDK_TEST_RAID=1 00:04:22.801 15:52:20 -- spdk_repo/autorun-spdk.conf@5 -- $ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:04:22.801 15:52:20 -- spdk_repo/autorun-spdk.conf@6 -- $ RUN_NIGHTLY=0 00:04:22.801 15:52:20 -- spdk/autorun.sh@22 -- $ trap 'timing_finish || exit 1' EXIT 00:04:22.801 15:52:20 -- spdk/autorun.sh@25 -- $ /home/vagrant/spdk_repo/spdk/autobuild.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:04:22.801 15:52:20 -- common/autotest_common.sh@1692 -- $ [[ n == y ]] 00:04:22.801 15:52:20 -- common/autobuild_common.sh@15 -- $ source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:04:22.801 15:52:20 -- scripts/common.sh@15 -- $ shopt -s extglob 00:04:22.801 15:52:20 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:04:22.801 15:52:20 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:04:22.801 15:52:20 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:04:22.801 15:52:20 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:22.801 15:52:20 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:22.801 15:52:20 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:22.801 15:52:20 -- paths/export.sh@5 -- $ export PATH 00:04:22.801 15:52:20 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:22.801 15:52:20 -- common/autobuild_common.sh@492 -- $ out=/home/vagrant/spdk_repo/spdk/../output 00:04:22.801 15:52:20 -- common/autobuild_common.sh@493 -- $ date +%s 00:04:22.801 15:52:20 -- common/autobuild_common.sh@493 -- $ mktemp -dt spdk_1732117940.XXXXXX 00:04:22.801 15:52:20 -- common/autobuild_common.sh@493 -- $ SPDK_WORKSPACE=/tmp/spdk_1732117940.bgzsXU 00:04:22.801 15:52:20 -- common/autobuild_common.sh@495 -- $ [[ -n '' ]] 00:04:22.801 15:52:20 -- common/autobuild_common.sh@499 -- $ '[' -n '' ']' 00:04:22.801 15:52:20 -- common/autobuild_common.sh@502 -- $ scanbuild_exclude='--exclude /home/vagrant/spdk_repo/spdk/dpdk/' 00:04:22.801 15:52:20 -- common/autobuild_common.sh@506 -- $ scanbuild_exclude+=' --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp' 00:04:22.801 15:52:20 -- common/autobuild_common.sh@508 -- $ scanbuild='scan-build -o /home/vagrant/spdk_repo/spdk/../output/scan-build-tmp --exclude /home/vagrant/spdk_repo/spdk/dpdk/ --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp --status-bugs' 00:04:22.801 15:52:20 -- common/autobuild_common.sh@509 -- $ get_config_params 00:04:22.801 15:52:20 -- common/autotest_common.sh@409 -- $ xtrace_disable 00:04:22.801 15:52:20 -- common/autotest_common.sh@10 -- $ set +x 00:04:22.801 15:52:20 -- common/autobuild_common.sh@509 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f' 00:04:22.801 15:52:20 -- common/autobuild_common.sh@511 -- $ start_monitor_resources 00:04:22.801 15:52:20 -- pm/common@17 -- $ local monitor 00:04:22.801 15:52:20 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:22.801 15:52:20 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:22.801 15:52:20 -- pm/common@25 -- $ sleep 1 00:04:22.801 15:52:20 -- pm/common@21 -- $ date +%s 00:04:22.801 15:52:20 -- pm/common@21 -- $ date +%s 00:04:22.801 15:52:20 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1732117940 00:04:22.801 15:52:20 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1732117940 00:04:22.801 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1732117940_collect-vmstat.pm.log 00:04:22.801 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1732117940_collect-cpu-load.pm.log 00:04:23.735 15:52:21 -- common/autobuild_common.sh@512 -- $ trap stop_monitor_resources EXIT 00:04:23.735 15:52:21 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:04:23.735 15:52:21 -- spdk/autobuild.sh@12 -- $ umask 022 00:04:23.735 15:52:21 -- spdk/autobuild.sh@13 -- $ cd /home/vagrant/spdk_repo/spdk 00:04:23.735 15:52:21 -- spdk/autobuild.sh@16 -- $ date -u 00:04:23.735 Wed Nov 20 03:52:21 PM UTC 2024 00:04:23.735 15:52:21 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:04:23.735 v25.01-pre-241-g0728de5b0 00:04:23.735 15:52:21 -- spdk/autobuild.sh@19 -- $ '[' 1 -eq 1 ']' 00:04:23.735 15:52:21 -- spdk/autobuild.sh@20 -- $ run_test asan echo 'using asan' 00:04:23.735 15:52:21 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:04:23.735 15:52:21 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:04:23.735 15:52:21 -- common/autotest_common.sh@10 -- $ set +x 00:04:23.735 ************************************ 00:04:23.735 START TEST asan 00:04:23.735 ************************************ 00:04:23.735 using asan 00:04:23.735 15:52:21 asan -- common/autotest_common.sh@1129 -- $ echo 'using asan' 00:04:23.735 00:04:23.735 real 0m0.000s 00:04:23.735 user 0m0.000s 00:04:23.735 sys 0m0.000s 00:04:23.735 15:52:21 asan -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:04:23.735 ************************************ 00:04:23.735 END TEST asan 00:04:23.735 ************************************ 00:04:23.735 15:52:21 asan -- common/autotest_common.sh@10 -- $ set +x 00:04:23.735 15:52:21 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:04:23.735 15:52:21 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:04:23.735 15:52:21 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:04:23.735 15:52:21 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:04:23.735 15:52:21 -- common/autotest_common.sh@10 -- $ set +x 00:04:23.735 ************************************ 00:04:23.735 START TEST ubsan 00:04:23.735 ************************************ 00:04:23.735 using ubsan 00:04:23.735 15:52:21 ubsan -- common/autotest_common.sh@1129 -- $ echo 'using ubsan' 00:04:23.735 00:04:23.735 real 0m0.000s 00:04:23.735 user 0m0.000s 00:04:23.735 sys 0m0.000s 00:04:23.735 ************************************ 00:04:23.735 END TEST ubsan 00:04:23.735 15:52:21 ubsan -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:04:23.735 15:52:21 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:04:23.735 ************************************ 00:04:23.735 15:52:21 -- spdk/autobuild.sh@27 -- $ '[' -n '' ']' 00:04:23.735 15:52:21 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:04:23.735 15:52:21 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:04:23.735 15:52:21 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:04:23.735 15:52:21 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:04:23.735 15:52:21 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:04:23.735 15:52:21 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:04:23.735 15:52:21 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:04:23.735 15:52:21 -- spdk/autobuild.sh@67 -- $ /home/vagrant/spdk_repo/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-shared 00:04:23.993 Using default SPDK env in /home/vagrant/spdk_repo/spdk/lib/env_dpdk 00:04:23.993 Using default DPDK in /home/vagrant/spdk_repo/spdk/dpdk/build 00:04:24.252 Using 'verbs' RDMA provider 00:04:35.149 Configuring ISA-L (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal.log)...done. 00:04:45.139 Configuring ISA-L-crypto (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal-crypto.log)...done. 00:04:45.139 Creating mk/config.mk...done. 00:04:45.139 Creating mk/cc.flags.mk...done. 00:04:45.139 Type 'make' to build. 00:04:45.139 15:52:43 -- spdk/autobuild.sh@70 -- $ run_test make make -j10 00:04:45.139 15:52:43 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:04:45.139 15:52:43 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:04:45.139 15:52:43 -- common/autotest_common.sh@10 -- $ set +x 00:04:45.139 ************************************ 00:04:45.139 START TEST make 00:04:45.139 ************************************ 00:04:45.139 15:52:43 make -- common/autotest_common.sh@1129 -- $ make -j10 00:04:45.397 make[1]: Nothing to be done for 'all'. 00:04:55.399 The Meson build system 00:04:55.399 Version: 1.5.0 00:04:55.399 Source dir: /home/vagrant/spdk_repo/spdk/dpdk 00:04:55.399 Build dir: /home/vagrant/spdk_repo/spdk/dpdk/build-tmp 00:04:55.399 Build type: native build 00:04:55.399 Program cat found: YES (/usr/bin/cat) 00:04:55.399 Project name: DPDK 00:04:55.399 Project version: 24.03.0 00:04:55.399 C compiler for the host machine: cc (gcc 13.3.1 "cc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:04:55.399 C linker for the host machine: cc ld.bfd 2.40-14 00:04:55.399 Host machine cpu family: x86_64 00:04:55.399 Host machine cpu: x86_64 00:04:55.399 Message: ## Building in Developer Mode ## 00:04:55.399 Program pkg-config found: YES (/usr/bin/pkg-config) 00:04:55.399 Program check-symbols.sh found: YES (/home/vagrant/spdk_repo/spdk/dpdk/buildtools/check-symbols.sh) 00:04:55.399 Program options-ibverbs-static.sh found: YES (/home/vagrant/spdk_repo/spdk/dpdk/buildtools/options-ibverbs-static.sh) 00:04:55.399 Program python3 found: YES (/usr/bin/python3) 00:04:55.399 Program cat found: YES (/usr/bin/cat) 00:04:55.399 Compiler for C supports arguments -march=native: YES 00:04:55.399 Checking for size of "void *" : 8 00:04:55.399 Checking for size of "void *" : 8 (cached) 00:04:55.399 Compiler for C supports link arguments -Wl,--undefined-version: YES 00:04:55.399 Library m found: YES 00:04:55.399 Library numa found: YES 00:04:55.399 Has header "numaif.h" : YES 00:04:55.399 Library fdt found: NO 00:04:55.399 Library execinfo found: NO 00:04:55.399 Has header "execinfo.h" : YES 00:04:55.399 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:04:55.399 Run-time dependency libarchive found: NO (tried pkgconfig) 00:04:55.399 Run-time dependency libbsd found: NO (tried pkgconfig) 00:04:55.399 Run-time dependency jansson found: NO (tried pkgconfig) 00:04:55.399 Run-time dependency openssl found: YES 3.1.1 00:04:55.399 Run-time dependency libpcap found: YES 1.10.4 00:04:55.399 Has header "pcap.h" with dependency libpcap: YES 00:04:55.399 Compiler for C supports arguments -Wcast-qual: YES 00:04:55.399 Compiler for C supports arguments -Wdeprecated: YES 00:04:55.399 Compiler for C supports arguments -Wformat: YES 00:04:55.399 Compiler for C supports arguments -Wformat-nonliteral: NO 00:04:55.399 Compiler for C supports arguments -Wformat-security: NO 00:04:55.399 Compiler for C supports arguments -Wmissing-declarations: YES 00:04:55.399 Compiler for C supports arguments -Wmissing-prototypes: YES 00:04:55.399 Compiler for C supports arguments -Wnested-externs: YES 00:04:55.399 Compiler for C supports arguments -Wold-style-definition: YES 00:04:55.399 Compiler for C supports arguments -Wpointer-arith: YES 00:04:55.399 Compiler for C supports arguments -Wsign-compare: YES 00:04:55.399 Compiler for C supports arguments -Wstrict-prototypes: YES 00:04:55.399 Compiler for C supports arguments -Wundef: YES 00:04:55.399 Compiler for C supports arguments -Wwrite-strings: YES 00:04:55.399 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:04:55.399 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:04:55.399 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:04:55.399 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:04:55.399 Program objdump found: YES (/usr/bin/objdump) 00:04:55.399 Compiler for C supports arguments -mavx512f: YES 00:04:55.399 Checking if "AVX512 checking" compiles: YES 00:04:55.399 Fetching value of define "__SSE4_2__" : 1 00:04:55.399 Fetching value of define "__AES__" : 1 00:04:55.399 Fetching value of define "__AVX__" : 1 00:04:55.399 Fetching value of define "__AVX2__" : 1 00:04:55.399 Fetching value of define "__AVX512BW__" : 1 00:04:55.399 Fetching value of define "__AVX512CD__" : 1 00:04:55.399 Fetching value of define "__AVX512DQ__" : 1 00:04:55.399 Fetching value of define "__AVX512F__" : 1 00:04:55.399 Fetching value of define "__AVX512VL__" : 1 00:04:55.399 Fetching value of define "__PCLMUL__" : 1 00:04:55.399 Fetching value of define "__RDRND__" : 1 00:04:55.399 Fetching value of define "__RDSEED__" : 1 00:04:55.399 Fetching value of define "__VPCLMULQDQ__" : 1 00:04:55.399 Fetching value of define "__znver1__" : (undefined) 00:04:55.399 Fetching value of define "__znver2__" : (undefined) 00:04:55.399 Fetching value of define "__znver3__" : (undefined) 00:04:55.399 Fetching value of define "__znver4__" : (undefined) 00:04:55.399 Library asan found: YES 00:04:55.399 Compiler for C supports arguments -Wno-format-truncation: YES 00:04:55.399 Message: lib/log: Defining dependency "log" 00:04:55.399 Message: lib/kvargs: Defining dependency "kvargs" 00:04:55.399 Message: lib/telemetry: Defining dependency "telemetry" 00:04:55.399 Library rt found: YES 00:04:55.399 Checking for function "getentropy" : NO 00:04:55.399 Message: lib/eal: Defining dependency "eal" 00:04:55.399 Message: lib/ring: Defining dependency "ring" 00:04:55.399 Message: lib/rcu: Defining dependency "rcu" 00:04:55.399 Message: lib/mempool: Defining dependency "mempool" 00:04:55.399 Message: lib/mbuf: Defining dependency "mbuf" 00:04:55.399 Fetching value of define "__PCLMUL__" : 1 (cached) 00:04:55.399 Fetching value of define "__AVX512F__" : 1 (cached) 00:04:55.399 Fetching value of define "__AVX512BW__" : 1 (cached) 00:04:55.399 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:04:55.399 Fetching value of define "__AVX512VL__" : 1 (cached) 00:04:55.399 Fetching value of define "__VPCLMULQDQ__" : 1 (cached) 00:04:55.399 Compiler for C supports arguments -mpclmul: YES 00:04:55.399 Compiler for C supports arguments -maes: YES 00:04:55.399 Compiler for C supports arguments -mavx512f: YES (cached) 00:04:55.399 Compiler for C supports arguments -mavx512bw: YES 00:04:55.399 Compiler for C supports arguments -mavx512dq: YES 00:04:55.399 Compiler for C supports arguments -mavx512vl: YES 00:04:55.399 Compiler for C supports arguments -mvpclmulqdq: YES 00:04:55.399 Compiler for C supports arguments -mavx2: YES 00:04:55.399 Compiler for C supports arguments -mavx: YES 00:04:55.399 Message: lib/net: Defining dependency "net" 00:04:55.399 Message: lib/meter: Defining dependency "meter" 00:04:55.399 Message: lib/ethdev: Defining dependency "ethdev" 00:04:55.399 Message: lib/pci: Defining dependency "pci" 00:04:55.399 Message: lib/cmdline: Defining dependency "cmdline" 00:04:55.399 Message: lib/hash: Defining dependency "hash" 00:04:55.399 Message: lib/timer: Defining dependency "timer" 00:04:55.399 Message: lib/compressdev: Defining dependency "compressdev" 00:04:55.399 Message: lib/cryptodev: Defining dependency "cryptodev" 00:04:55.399 Message: lib/dmadev: Defining dependency "dmadev" 00:04:55.399 Compiler for C supports arguments -Wno-cast-qual: YES 00:04:55.399 Message: lib/power: Defining dependency "power" 00:04:55.399 Message: lib/reorder: Defining dependency "reorder" 00:04:55.399 Message: lib/security: Defining dependency "security" 00:04:55.399 Has header "linux/userfaultfd.h" : YES 00:04:55.399 Has header "linux/vduse.h" : YES 00:04:55.399 Message: lib/vhost: Defining dependency "vhost" 00:04:55.399 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:04:55.399 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:04:55.399 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:04:55.399 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:04:55.399 Message: Disabling raw/* drivers: missing internal dependency "rawdev" 00:04:55.399 Message: Disabling regex/* drivers: missing internal dependency "regexdev" 00:04:55.399 Message: Disabling ml/* drivers: missing internal dependency "mldev" 00:04:55.399 Message: Disabling event/* drivers: missing internal dependency "eventdev" 00:04:55.399 Message: Disabling baseband/* drivers: missing internal dependency "bbdev" 00:04:55.399 Message: Disabling gpu/* drivers: missing internal dependency "gpudev" 00:04:55.399 Program doxygen found: YES (/usr/local/bin/doxygen) 00:04:55.399 Configuring doxy-api-html.conf using configuration 00:04:55.399 Configuring doxy-api-man.conf using configuration 00:04:55.399 Program mandb found: YES (/usr/bin/mandb) 00:04:55.399 Program sphinx-build found: NO 00:04:55.399 Configuring rte_build_config.h using configuration 00:04:55.399 Message: 00:04:55.399 ================= 00:04:55.399 Applications Enabled 00:04:55.399 ================= 00:04:55.399 00:04:55.399 apps: 00:04:55.399 00:04:55.399 00:04:55.399 Message: 00:04:55.399 ================= 00:04:55.399 Libraries Enabled 00:04:55.399 ================= 00:04:55.399 00:04:55.399 libs: 00:04:55.399 log, kvargs, telemetry, eal, ring, rcu, mempool, mbuf, 00:04:55.399 net, meter, ethdev, pci, cmdline, hash, timer, compressdev, 00:04:55.399 cryptodev, dmadev, power, reorder, security, vhost, 00:04:55.399 00:04:55.399 Message: 00:04:55.399 =============== 00:04:55.399 Drivers Enabled 00:04:55.399 =============== 00:04:55.399 00:04:55.399 common: 00:04:55.399 00:04:55.399 bus: 00:04:55.399 pci, vdev, 00:04:55.399 mempool: 00:04:55.399 ring, 00:04:55.399 dma: 00:04:55.399 00:04:55.399 net: 00:04:55.399 00:04:55.399 crypto: 00:04:55.399 00:04:55.399 compress: 00:04:55.399 00:04:55.399 vdpa: 00:04:55.399 00:04:55.399 00:04:55.399 Message: 00:04:55.399 ================= 00:04:55.399 Content Skipped 00:04:55.399 ================= 00:04:55.399 00:04:55.399 apps: 00:04:55.399 dumpcap: explicitly disabled via build config 00:04:55.399 graph: explicitly disabled via build config 00:04:55.399 pdump: explicitly disabled via build config 00:04:55.399 proc-info: explicitly disabled via build config 00:04:55.399 test-acl: explicitly disabled via build config 00:04:55.399 test-bbdev: explicitly disabled via build config 00:04:55.399 test-cmdline: explicitly disabled via build config 00:04:55.399 test-compress-perf: explicitly disabled via build config 00:04:55.399 test-crypto-perf: explicitly disabled via build config 00:04:55.400 test-dma-perf: explicitly disabled via build config 00:04:55.400 test-eventdev: explicitly disabled via build config 00:04:55.400 test-fib: explicitly disabled via build config 00:04:55.400 test-flow-perf: explicitly disabled via build config 00:04:55.400 test-gpudev: explicitly disabled via build config 00:04:55.400 test-mldev: explicitly disabled via build config 00:04:55.400 test-pipeline: explicitly disabled via build config 00:04:55.400 test-pmd: explicitly disabled via build config 00:04:55.400 test-regex: explicitly disabled via build config 00:04:55.400 test-sad: explicitly disabled via build config 00:04:55.400 test-security-perf: explicitly disabled via build config 00:04:55.400 00:04:55.400 libs: 00:04:55.400 argparse: explicitly disabled via build config 00:04:55.400 metrics: explicitly disabled via build config 00:04:55.400 acl: explicitly disabled via build config 00:04:55.400 bbdev: explicitly disabled via build config 00:04:55.400 bitratestats: explicitly disabled via build config 00:04:55.400 bpf: explicitly disabled via build config 00:04:55.400 cfgfile: explicitly disabled via build config 00:04:55.400 distributor: explicitly disabled via build config 00:04:55.400 efd: explicitly disabled via build config 00:04:55.400 eventdev: explicitly disabled via build config 00:04:55.400 dispatcher: explicitly disabled via build config 00:04:55.400 gpudev: explicitly disabled via build config 00:04:55.400 gro: explicitly disabled via build config 00:04:55.400 gso: explicitly disabled via build config 00:04:55.400 ip_frag: explicitly disabled via build config 00:04:55.400 jobstats: explicitly disabled via build config 00:04:55.400 latencystats: explicitly disabled via build config 00:04:55.400 lpm: explicitly disabled via build config 00:04:55.400 member: explicitly disabled via build config 00:04:55.400 pcapng: explicitly disabled via build config 00:04:55.400 rawdev: explicitly disabled via build config 00:04:55.400 regexdev: explicitly disabled via build config 00:04:55.400 mldev: explicitly disabled via build config 00:04:55.400 rib: explicitly disabled via build config 00:04:55.400 sched: explicitly disabled via build config 00:04:55.400 stack: explicitly disabled via build config 00:04:55.400 ipsec: explicitly disabled via build config 00:04:55.400 pdcp: explicitly disabled via build config 00:04:55.400 fib: explicitly disabled via build config 00:04:55.400 port: explicitly disabled via build config 00:04:55.400 pdump: explicitly disabled via build config 00:04:55.400 table: explicitly disabled via build config 00:04:55.400 pipeline: explicitly disabled via build config 00:04:55.400 graph: explicitly disabled via build config 00:04:55.400 node: explicitly disabled via build config 00:04:55.400 00:04:55.400 drivers: 00:04:55.400 common/cpt: not in enabled drivers build config 00:04:55.400 common/dpaax: not in enabled drivers build config 00:04:55.400 common/iavf: not in enabled drivers build config 00:04:55.400 common/idpf: not in enabled drivers build config 00:04:55.400 common/ionic: not in enabled drivers build config 00:04:55.400 common/mvep: not in enabled drivers build config 00:04:55.400 common/octeontx: not in enabled drivers build config 00:04:55.400 bus/auxiliary: not in enabled drivers build config 00:04:55.400 bus/cdx: not in enabled drivers build config 00:04:55.400 bus/dpaa: not in enabled drivers build config 00:04:55.400 bus/fslmc: not in enabled drivers build config 00:04:55.400 bus/ifpga: not in enabled drivers build config 00:04:55.400 bus/platform: not in enabled drivers build config 00:04:55.400 bus/uacce: not in enabled drivers build config 00:04:55.400 bus/vmbus: not in enabled drivers build config 00:04:55.400 common/cnxk: not in enabled drivers build config 00:04:55.400 common/mlx5: not in enabled drivers build config 00:04:55.400 common/nfp: not in enabled drivers build config 00:04:55.400 common/nitrox: not in enabled drivers build config 00:04:55.400 common/qat: not in enabled drivers build config 00:04:55.400 common/sfc_efx: not in enabled drivers build config 00:04:55.400 mempool/bucket: not in enabled drivers build config 00:04:55.400 mempool/cnxk: not in enabled drivers build config 00:04:55.400 mempool/dpaa: not in enabled drivers build config 00:04:55.400 mempool/dpaa2: not in enabled drivers build config 00:04:55.400 mempool/octeontx: not in enabled drivers build config 00:04:55.400 mempool/stack: not in enabled drivers build config 00:04:55.400 dma/cnxk: not in enabled drivers build config 00:04:55.400 dma/dpaa: not in enabled drivers build config 00:04:55.400 dma/dpaa2: not in enabled drivers build config 00:04:55.400 dma/hisilicon: not in enabled drivers build config 00:04:55.400 dma/idxd: not in enabled drivers build config 00:04:55.400 dma/ioat: not in enabled drivers build config 00:04:55.400 dma/skeleton: not in enabled drivers build config 00:04:55.400 net/af_packet: not in enabled drivers build config 00:04:55.400 net/af_xdp: not in enabled drivers build config 00:04:55.400 net/ark: not in enabled drivers build config 00:04:55.400 net/atlantic: not in enabled drivers build config 00:04:55.400 net/avp: not in enabled drivers build config 00:04:55.400 net/axgbe: not in enabled drivers build config 00:04:55.400 net/bnx2x: not in enabled drivers build config 00:04:55.400 net/bnxt: not in enabled drivers build config 00:04:55.400 net/bonding: not in enabled drivers build config 00:04:55.400 net/cnxk: not in enabled drivers build config 00:04:55.400 net/cpfl: not in enabled drivers build config 00:04:55.400 net/cxgbe: not in enabled drivers build config 00:04:55.400 net/dpaa: not in enabled drivers build config 00:04:55.400 net/dpaa2: not in enabled drivers build config 00:04:55.400 net/e1000: not in enabled drivers build config 00:04:55.400 net/ena: not in enabled drivers build config 00:04:55.400 net/enetc: not in enabled drivers build config 00:04:55.400 net/enetfec: not in enabled drivers build config 00:04:55.400 net/enic: not in enabled drivers build config 00:04:55.400 net/failsafe: not in enabled drivers build config 00:04:55.400 net/fm10k: not in enabled drivers build config 00:04:55.400 net/gve: not in enabled drivers build config 00:04:55.400 net/hinic: not in enabled drivers build config 00:04:55.400 net/hns3: not in enabled drivers build config 00:04:55.400 net/i40e: not in enabled drivers build config 00:04:55.400 net/iavf: not in enabled drivers build config 00:04:55.400 net/ice: not in enabled drivers build config 00:04:55.400 net/idpf: not in enabled drivers build config 00:04:55.400 net/igc: not in enabled drivers build config 00:04:55.400 net/ionic: not in enabled drivers build config 00:04:55.400 net/ipn3ke: not in enabled drivers build config 00:04:55.400 net/ixgbe: not in enabled drivers build config 00:04:55.400 net/mana: not in enabled drivers build config 00:04:55.400 net/memif: not in enabled drivers build config 00:04:55.400 net/mlx4: not in enabled drivers build config 00:04:55.400 net/mlx5: not in enabled drivers build config 00:04:55.400 net/mvneta: not in enabled drivers build config 00:04:55.400 net/mvpp2: not in enabled drivers build config 00:04:55.400 net/netvsc: not in enabled drivers build config 00:04:55.400 net/nfb: not in enabled drivers build config 00:04:55.400 net/nfp: not in enabled drivers build config 00:04:55.400 net/ngbe: not in enabled drivers build config 00:04:55.401 net/null: not in enabled drivers build config 00:04:55.401 net/octeontx: not in enabled drivers build config 00:04:55.401 net/octeon_ep: not in enabled drivers build config 00:04:55.401 net/pcap: not in enabled drivers build config 00:04:55.401 net/pfe: not in enabled drivers build config 00:04:55.401 net/qede: not in enabled drivers build config 00:04:55.401 net/ring: not in enabled drivers build config 00:04:55.401 net/sfc: not in enabled drivers build config 00:04:55.401 net/softnic: not in enabled drivers build config 00:04:55.401 net/tap: not in enabled drivers build config 00:04:55.401 net/thunderx: not in enabled drivers build config 00:04:55.401 net/txgbe: not in enabled drivers build config 00:04:55.401 net/vdev_netvsc: not in enabled drivers build config 00:04:55.401 net/vhost: not in enabled drivers build config 00:04:55.401 net/virtio: not in enabled drivers build config 00:04:55.401 net/vmxnet3: not in enabled drivers build config 00:04:55.401 raw/*: missing internal dependency, "rawdev" 00:04:55.401 crypto/armv8: not in enabled drivers build config 00:04:55.401 crypto/bcmfs: not in enabled drivers build config 00:04:55.401 crypto/caam_jr: not in enabled drivers build config 00:04:55.401 crypto/ccp: not in enabled drivers build config 00:04:55.401 crypto/cnxk: not in enabled drivers build config 00:04:55.401 crypto/dpaa_sec: not in enabled drivers build config 00:04:55.401 crypto/dpaa2_sec: not in enabled drivers build config 00:04:55.401 crypto/ipsec_mb: not in enabled drivers build config 00:04:55.401 crypto/mlx5: not in enabled drivers build config 00:04:55.401 crypto/mvsam: not in enabled drivers build config 00:04:55.401 crypto/nitrox: not in enabled drivers build config 00:04:55.401 crypto/null: not in enabled drivers build config 00:04:55.401 crypto/octeontx: not in enabled drivers build config 00:04:55.401 crypto/openssl: not in enabled drivers build config 00:04:55.401 crypto/scheduler: not in enabled drivers build config 00:04:55.401 crypto/uadk: not in enabled drivers build config 00:04:55.401 crypto/virtio: not in enabled drivers build config 00:04:55.401 compress/isal: not in enabled drivers build config 00:04:55.401 compress/mlx5: not in enabled drivers build config 00:04:55.401 compress/nitrox: not in enabled drivers build config 00:04:55.401 compress/octeontx: not in enabled drivers build config 00:04:55.401 compress/zlib: not in enabled drivers build config 00:04:55.401 regex/*: missing internal dependency, "regexdev" 00:04:55.401 ml/*: missing internal dependency, "mldev" 00:04:55.401 vdpa/ifc: not in enabled drivers build config 00:04:55.401 vdpa/mlx5: not in enabled drivers build config 00:04:55.401 vdpa/nfp: not in enabled drivers build config 00:04:55.401 vdpa/sfc: not in enabled drivers build config 00:04:55.401 event/*: missing internal dependency, "eventdev" 00:04:55.401 baseband/*: missing internal dependency, "bbdev" 00:04:55.401 gpu/*: missing internal dependency, "gpudev" 00:04:55.401 00:04:55.401 00:04:55.401 Build targets in project: 84 00:04:55.401 00:04:55.401 DPDK 24.03.0 00:04:55.401 00:04:55.401 User defined options 00:04:55.401 buildtype : debug 00:04:55.401 default_library : shared 00:04:55.401 libdir : lib 00:04:55.401 prefix : /home/vagrant/spdk_repo/spdk/dpdk/build 00:04:55.401 b_sanitize : address 00:04:55.401 c_args : -Wno-stringop-overflow -fcommon -Wno-stringop-overread -Wno-array-bounds -fPIC -Werror 00:04:55.401 c_link_args : 00:04:55.401 cpu_instruction_set: native 00:04:55.401 disable_apps : dumpcap,graph,pdump,proc-info,test-acl,test-bbdev,test-cmdline,test-compress-perf,test-crypto-perf,test-dma-perf,test-eventdev,test-fib,test-flow-perf,test-gpudev,test-mldev,test-pipeline,test-pmd,test-regex,test-sad,test-security-perf,test 00:04:55.401 disable_libs : acl,argparse,bbdev,bitratestats,bpf,cfgfile,dispatcher,distributor,efd,eventdev,fib,gpudev,graph,gro,gso,ip_frag,ipsec,jobstats,latencystats,lpm,member,metrics,mldev,node,pcapng,pdcp,pdump,pipeline,port,rawdev,regexdev,rib,sched,stack,table 00:04:55.401 enable_docs : false 00:04:55.401 enable_drivers : bus,bus/pci,bus/vdev,mempool/ring,power/acpi,power/amd_pstate,power/cppc,power/intel_pstate,power/intel_uncore,power/kvm_vm 00:04:55.401 enable_kmods : false 00:04:55.401 max_lcores : 128 00:04:55.401 tests : false 00:04:55.401 00:04:55.401 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:04:55.401 ninja: Entering directory `/home/vagrant/spdk_repo/spdk/dpdk/build-tmp' 00:04:55.401 [1/267] Compiling C object lib/librte_log.a.p/log_log_linux.c.o 00:04:55.401 [2/267] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:04:55.401 [3/267] Linking static target lib/librte_kvargs.a 00:04:55.401 [4/267] Compiling C object lib/librte_log.a.p/log_log.c.o 00:04:55.401 [5/267] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:04:55.401 [6/267] Linking static target lib/librte_log.a 00:04:55.401 [7/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:04:55.401 [8/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:04:55.401 [9/267] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:04:55.401 [10/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:04:55.401 [11/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:04:55.401 [12/267] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:04:55.401 [13/267] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:04:55.401 [14/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:04:55.401 [15/267] Linking static target lib/librte_telemetry.a 00:04:55.401 [16/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:04:55.658 [17/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:04:55.658 [18/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:04:55.916 [19/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:04:55.916 [20/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:04:55.916 [21/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:04:55.916 [22/267] Generating lib/log.sym_chk with a custom command (wrapped by meson to capture output) 00:04:55.916 [23/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:04:55.916 [24/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:04:55.916 [25/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:04:55.916 [26/267] Linking target lib/librte_log.so.24.1 00:04:56.173 [27/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:04:56.173 [28/267] Generating symbol file lib/librte_log.so.24.1.p/librte_log.so.24.1.symbols 00:04:56.173 [29/267] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:04:56.173 [30/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:04:56.173 [31/267] Linking target lib/librte_kvargs.so.24.1 00:04:56.173 [32/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:04:56.173 [33/267] Linking target lib/librte_telemetry.so.24.1 00:04:56.173 [34/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:04:56.173 [35/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:04:56.173 [36/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:04:56.430 [37/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:04:56.430 [38/267] Generating symbol file lib/librte_kvargs.so.24.1.p/librte_kvargs.so.24.1.symbols 00:04:56.430 [39/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:04:56.430 [40/267] Generating symbol file lib/librte_telemetry.so.24.1.p/librte_telemetry.so.24.1.symbols 00:04:56.430 [41/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:04:56.430 [42/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:04:56.430 [43/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:04:56.430 [44/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:04:56.687 [45/267] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:04:56.687 [46/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:04:56.687 [47/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:04:56.688 [48/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:04:56.688 [49/267] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:04:56.945 [50/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:04:56.945 [51/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:04:56.945 [52/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:04:56.945 [53/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:04:56.945 [54/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:04:56.945 [55/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:04:56.945 [56/267] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:04:57.202 [57/267] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:04:57.202 [58/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:04:57.202 [59/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:04:57.202 [60/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:04:57.202 [61/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:04:57.202 [62/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:04:57.202 [63/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:04:57.459 [64/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:04:57.459 [65/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:04:57.459 [66/267] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:04:57.459 [67/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:04:57.459 [68/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:04:57.459 [69/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:04:57.716 [70/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:04:57.716 [71/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:04:57.716 [72/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:04:57.716 [73/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:04:57.716 [74/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:04:57.716 [75/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:04:57.716 [76/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:04:57.716 [77/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:04:57.973 [78/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:04:57.973 [79/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:04:57.973 [80/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:04:57.973 [81/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:04:58.230 [82/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:04:58.230 [83/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:04:58.231 [84/267] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:04:58.231 [85/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:04:58.231 [86/267] Linking static target lib/librte_ring.a 00:04:58.231 [87/267] Linking static target lib/librte_eal.a 00:04:58.231 [88/267] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:04:58.231 [89/267] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:04:58.231 [90/267] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:04:58.231 [91/267] Linking static target lib/librte_rcu.a 00:04:58.488 [92/267] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:04:58.488 [93/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:04:58.488 [94/267] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:04:58.488 [95/267] Linking static target lib/librte_mempool.a 00:04:58.488 [96/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:04:58.488 [97/267] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:04:58.746 [98/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:04:58.746 [99/267] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:04:58.747 [100/267] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:04:58.747 [101/267] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:04:58.747 [102/267] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:04:58.747 [103/267] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:04:59.003 [104/267] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:04:59.003 [105/267] Linking static target lib/librte_meter.a 00:04:59.003 [106/267] Compiling C object lib/librte_net.a.p/net_net_crc_avx512.c.o 00:04:59.003 [107/267] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:04:59.003 [108/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:04:59.003 [109/267] Linking static target lib/librte_net.a 00:04:59.003 [110/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:04:59.260 [111/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:04:59.260 [112/267] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:04:59.260 [113/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:04:59.260 [114/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:04:59.260 [115/267] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:04:59.260 [116/267] Linking static target lib/librte_mbuf.a 00:04:59.518 [117/267] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:04:59.518 [118/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:04:59.518 [119/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:04:59.518 [120/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_telemetry.c.o 00:04:59.776 [121/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:04:59.776 [122/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:04:59.776 [123/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:05:00.034 [124/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:05:00.034 [125/267] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:05:00.034 [126/267] Linking static target lib/librte_pci.a 00:05:00.034 [127/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:05:00.034 [128/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:05:00.034 [129/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:05:00.034 [130/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:05:00.034 [131/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:05:00.348 [132/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:05:00.348 [133/267] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:05:00.348 [134/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_linux_ethtool.c.o 00:05:00.348 [135/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:05:00.348 [136/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:05:00.348 [137/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:05:00.348 [138/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:05:00.348 [139/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:05:00.348 [140/267] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:05:00.348 [141/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:05:00.348 [142/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:05:00.348 [143/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:05:00.348 [144/267] Compiling C object lib/librte_hash.a.p/hash_rte_hash_crc.c.o 00:05:00.348 [145/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:05:00.734 [146/267] Linking static target lib/librte_cmdline.a 00:05:00.734 [147/267] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:05:00.734 [148/267] Linking static target lib/librte_timer.a 00:05:00.734 [149/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:05:00.734 [150/267] Compiling C object lib/librte_hash.a.p/hash_rte_thash_gfni.c.o 00:05:00.734 [151/267] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:05:00.734 [152/267] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:05:00.992 [153/267] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:05:00.992 [154/267] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:05:00.992 [155/267] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:05:01.249 [156/267] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:05:01.249 [157/267] Linking static target lib/librte_compressdev.a 00:05:01.249 [158/267] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:05:01.249 [159/267] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:05:01.249 [160/267] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:05:01.249 [161/267] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:05:01.249 [162/267] Linking static target lib/librte_hash.a 00:05:01.249 [163/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:05:01.506 [164/267] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev_trace_points.c.o 00:05:01.506 [165/267] Linking static target lib/librte_ethdev.a 00:05:01.506 [166/267] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:05:01.506 [167/267] Linking static target lib/librte_dmadev.a 00:05:01.506 [168/267] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:05:01.506 [169/267] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:05:01.506 [170/267] Compiling C object lib/librte_power.a.p/power_power_amd_pstate_cpufreq.c.o 00:05:01.506 [171/267] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:05:01.764 [172/267] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:05:01.764 [173/267] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:01.764 [174/267] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:05:02.021 [175/267] Compiling C object lib/librte_power.a.p/power_power_intel_uncore.c.o 00:05:02.021 [176/267] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:05:02.021 [177/267] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:05:02.021 [178/267] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:02.021 [179/267] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:05:02.021 [180/267] Compiling C object lib/librte_power.a.p/power_rte_power_uncore.c.o 00:05:02.021 [181/267] Linking static target lib/librte_cryptodev.a 00:05:02.021 [182/267] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:05:02.279 [183/267] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:05:02.279 [184/267] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:05:02.279 [185/267] Linking static target lib/librte_power.a 00:05:02.279 [186/267] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:05:02.536 [187/267] Linking static target lib/librte_reorder.a 00:05:02.536 [188/267] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:05:02.536 [189/267] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:05:02.536 [190/267] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:05:02.536 [191/267] Linking static target lib/librte_security.a 00:05:02.536 [192/267] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:05:02.794 [193/267] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:05:02.794 [194/267] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:05:03.052 [195/267] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:05:03.052 [196/267] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:05:03.052 [197/267] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net_ctrl.c.o 00:05:03.310 [198/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:05:03.310 [199/267] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:05:03.310 [200/267] Compiling C object lib/librte_vhost.a.p/vhost_vduse.c.o 00:05:03.310 [201/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:05:03.567 [202/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:05:03.567 [203/267] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:05:03.567 [204/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:05:03.567 [205/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:05:03.824 [206/267] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:05:03.824 [207/267] Linking static target drivers/libtmp_rte_bus_vdev.a 00:05:03.824 [208/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:05:03.824 [209/267] Linking static target drivers/libtmp_rte_bus_pci.a 00:05:03.824 [210/267] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:04.080 [211/267] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:05:04.080 [212/267] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:05:04.080 [213/267] Compiling C object drivers/librte_bus_vdev.so.24.1.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:05:04.080 [214/267] Linking static target drivers/librte_bus_vdev.a 00:05:04.080 [215/267] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:05:04.080 [216/267] Linking static target drivers/libtmp_rte_mempool_ring.a 00:05:04.080 [217/267] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:05:04.080 [218/267] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:05:04.080 [219/267] Compiling C object drivers/librte_bus_pci.so.24.1.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:05:04.081 [220/267] Linking static target drivers/librte_bus_pci.a 00:05:04.339 [221/267] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:04.339 [222/267] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:05:04.339 [223/267] Compiling C object drivers/librte_mempool_ring.so.24.1.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:05:04.339 [224/267] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:05:04.339 [225/267] Linking static target drivers/librte_mempool_ring.a 00:05:04.596 [226/267] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:05:04.596 [227/267] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:05:05.969 [228/267] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:05:05.969 [229/267] Linking target lib/librte_eal.so.24.1 00:05:05.969 [230/267] Generating symbol file lib/librte_eal.so.24.1.p/librte_eal.so.24.1.symbols 00:05:05.969 [231/267] Linking target lib/librte_ring.so.24.1 00:05:05.969 [232/267] Linking target lib/librte_meter.so.24.1 00:05:05.969 [233/267] Linking target lib/librte_timer.so.24.1 00:05:05.969 [234/267] Linking target lib/librte_dmadev.so.24.1 00:05:05.969 [235/267] Linking target drivers/librte_bus_vdev.so.24.1 00:05:05.969 [236/267] Linking target lib/librte_pci.so.24.1 00:05:05.969 [237/267] Generating symbol file lib/librte_ring.so.24.1.p/librte_ring.so.24.1.symbols 00:05:05.969 [238/267] Generating symbol file lib/librte_meter.so.24.1.p/librte_meter.so.24.1.symbols 00:05:05.969 [239/267] Linking target lib/librte_mempool.so.24.1 00:05:05.969 [240/267] Linking target lib/librte_rcu.so.24.1 00:05:05.969 [241/267] Generating symbol file lib/librte_pci.so.24.1.p/librte_pci.so.24.1.symbols 00:05:06.226 [242/267] Generating symbol file lib/librte_timer.so.24.1.p/librte_timer.so.24.1.symbols 00:05:06.226 [243/267] Generating symbol file lib/librte_dmadev.so.24.1.p/librte_dmadev.so.24.1.symbols 00:05:06.226 [244/267] Linking target drivers/librte_bus_pci.so.24.1 00:05:06.226 [245/267] Generating symbol file lib/librte_rcu.so.24.1.p/librte_rcu.so.24.1.symbols 00:05:06.226 [246/267] Generating symbol file lib/librte_mempool.so.24.1.p/librte_mempool.so.24.1.symbols 00:05:06.226 [247/267] Linking target drivers/librte_mempool_ring.so.24.1 00:05:06.226 [248/267] Linking target lib/librte_mbuf.so.24.1 00:05:06.226 [249/267] Generating symbol file lib/librte_mbuf.so.24.1.p/librte_mbuf.so.24.1.symbols 00:05:06.484 [250/267] Linking target lib/librte_compressdev.so.24.1 00:05:06.484 [251/267] Linking target lib/librte_cryptodev.so.24.1 00:05:06.484 [252/267] Linking target lib/librte_reorder.so.24.1 00:05:06.484 [253/267] Linking target lib/librte_net.so.24.1 00:05:06.484 [254/267] Generating symbol file lib/librte_cryptodev.so.24.1.p/librte_cryptodev.so.24.1.symbols 00:05:06.484 [255/267] Generating symbol file lib/librte_net.so.24.1.p/librte_net.so.24.1.symbols 00:05:06.484 [256/267] Linking target lib/librte_hash.so.24.1 00:05:06.484 [257/267] Linking target lib/librte_cmdline.so.24.1 00:05:06.484 [258/267] Linking target lib/librte_security.so.24.1 00:05:06.741 [259/267] Generating symbol file lib/librte_hash.so.24.1.p/librte_hash.so.24.1.symbols 00:05:06.998 [260/267] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:06.998 [261/267] Linking target lib/librte_ethdev.so.24.1 00:05:06.998 [262/267] Generating symbol file lib/librte_ethdev.so.24.1.p/librte_ethdev.so.24.1.symbols 00:05:07.256 [263/267] Linking target lib/librte_power.so.24.1 00:05:07.512 [264/267] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:05:07.512 [265/267] Linking static target lib/librte_vhost.a 00:05:08.884 [266/267] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:05:08.884 [267/267] Linking target lib/librte_vhost.so.24.1 00:05:08.884 INFO: autodetecting backend as ninja 00:05:08.884 INFO: calculating backend command to run: /usr/local/bin/ninja -C /home/vagrant/spdk_repo/spdk/dpdk/build-tmp -j 10 00:05:23.858 CC lib/log/log_flags.o 00:05:23.858 CC lib/log/log.o 00:05:23.858 CC lib/ut_mock/mock.o 00:05:23.858 CC lib/log/log_deprecated.o 00:05:23.858 CC lib/ut/ut.o 00:05:24.118 LIB libspdk_ut.a 00:05:24.118 LIB libspdk_log.a 00:05:24.118 SO libspdk_ut.so.2.0 00:05:24.118 LIB libspdk_ut_mock.a 00:05:24.118 SO libspdk_log.so.7.1 00:05:24.118 SO libspdk_ut_mock.so.6.0 00:05:24.118 SYMLINK libspdk_ut.so 00:05:24.118 SYMLINK libspdk_log.so 00:05:24.376 SYMLINK libspdk_ut_mock.so 00:05:24.376 CXX lib/trace_parser/trace.o 00:05:24.376 CC lib/util/base64.o 00:05:24.376 CC lib/util/bit_array.o 00:05:24.376 CC lib/util/cpuset.o 00:05:24.376 CC lib/util/crc16.o 00:05:24.376 CC lib/ioat/ioat.o 00:05:24.376 CC lib/util/crc32.o 00:05:24.376 CC lib/util/crc32c.o 00:05:24.376 CC lib/dma/dma.o 00:05:24.376 CC lib/vfio_user/host/vfio_user_pci.o 00:05:24.376 CC lib/util/crc32_ieee.o 00:05:24.376 CC lib/util/crc64.o 00:05:24.633 CC lib/util/dif.o 00:05:24.633 CC lib/util/fd.o 00:05:24.633 CC lib/util/fd_group.o 00:05:24.633 CC lib/util/file.o 00:05:24.633 LIB libspdk_dma.a 00:05:24.633 SO libspdk_dma.so.5.0 00:05:24.633 CC lib/util/hexlify.o 00:05:24.633 CC lib/vfio_user/host/vfio_user.o 00:05:24.633 LIB libspdk_ioat.a 00:05:24.633 SYMLINK libspdk_dma.so 00:05:24.633 CC lib/util/iov.o 00:05:24.633 SO libspdk_ioat.so.7.0 00:05:24.633 CC lib/util/math.o 00:05:24.633 CC lib/util/net.o 00:05:24.633 CC lib/util/pipe.o 00:05:24.633 SYMLINK libspdk_ioat.so 00:05:24.633 CC lib/util/strerror_tls.o 00:05:24.633 CC lib/util/string.o 00:05:24.890 LIB libspdk_vfio_user.a 00:05:24.890 SO libspdk_vfio_user.so.5.0 00:05:24.890 CC lib/util/uuid.o 00:05:24.890 CC lib/util/xor.o 00:05:24.890 CC lib/util/zipf.o 00:05:24.890 CC lib/util/md5.o 00:05:24.890 SYMLINK libspdk_vfio_user.so 00:05:25.147 LIB libspdk_util.a 00:05:25.147 SO libspdk_util.so.10.1 00:05:25.147 LIB libspdk_trace_parser.a 00:05:25.405 SO libspdk_trace_parser.so.6.0 00:05:25.405 SYMLINK libspdk_util.so 00:05:25.405 SYMLINK libspdk_trace_parser.so 00:05:25.405 CC lib/idxd/idxd.o 00:05:25.405 CC lib/conf/conf.o 00:05:25.405 CC lib/rdma_utils/rdma_utils.o 00:05:25.405 CC lib/idxd/idxd_user.o 00:05:25.405 CC lib/idxd/idxd_kernel.o 00:05:25.405 CC lib/json/json_parse.o 00:05:25.405 CC lib/json/json_util.o 00:05:25.405 CC lib/json/json_write.o 00:05:25.405 CC lib/env_dpdk/env.o 00:05:25.405 CC lib/vmd/vmd.o 00:05:25.662 CC lib/vmd/led.o 00:05:25.662 LIB libspdk_conf.a 00:05:25.662 CC lib/env_dpdk/memory.o 00:05:25.662 SO libspdk_conf.so.6.0 00:05:25.662 CC lib/env_dpdk/pci.o 00:05:25.662 CC lib/env_dpdk/init.o 00:05:25.662 CC lib/env_dpdk/threads.o 00:05:25.662 SYMLINK libspdk_conf.so 00:05:25.662 LIB libspdk_json.a 00:05:25.662 CC lib/env_dpdk/pci_ioat.o 00:05:25.662 LIB libspdk_rdma_utils.a 00:05:25.662 SO libspdk_rdma_utils.so.1.0 00:05:25.662 SO libspdk_json.so.6.0 00:05:25.920 SYMLINK libspdk_json.so 00:05:25.920 SYMLINK libspdk_rdma_utils.so 00:05:25.920 CC lib/env_dpdk/pci_virtio.o 00:05:25.920 CC lib/env_dpdk/pci_vmd.o 00:05:25.920 CC lib/env_dpdk/pci_idxd.o 00:05:25.920 CC lib/jsonrpc/jsonrpc_server.o 00:05:25.920 CC lib/env_dpdk/pci_event.o 00:05:25.920 CC lib/rdma_provider/common.o 00:05:26.179 CC lib/env_dpdk/sigbus_handler.o 00:05:26.179 CC lib/env_dpdk/pci_dpdk.o 00:05:26.179 CC lib/env_dpdk/pci_dpdk_2207.o 00:05:26.179 LIB libspdk_idxd.a 00:05:26.179 CC lib/env_dpdk/pci_dpdk_2211.o 00:05:26.179 SO libspdk_idxd.so.12.1 00:05:26.179 LIB libspdk_vmd.a 00:05:26.179 CC lib/rdma_provider/rdma_provider_verbs.o 00:05:26.179 SO libspdk_vmd.so.6.0 00:05:26.179 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:05:26.179 SYMLINK libspdk_idxd.so 00:05:26.179 CC lib/jsonrpc/jsonrpc_client.o 00:05:26.179 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:05:26.179 SYMLINK libspdk_vmd.so 00:05:26.440 LIB libspdk_rdma_provider.a 00:05:26.440 SO libspdk_rdma_provider.so.7.0 00:05:26.440 SYMLINK libspdk_rdma_provider.so 00:05:26.440 LIB libspdk_jsonrpc.a 00:05:26.440 SO libspdk_jsonrpc.so.6.0 00:05:26.440 SYMLINK libspdk_jsonrpc.so 00:05:26.697 CC lib/rpc/rpc.o 00:05:26.956 LIB libspdk_env_dpdk.a 00:05:26.956 LIB libspdk_rpc.a 00:05:26.956 SO libspdk_env_dpdk.so.15.1 00:05:26.956 SO libspdk_rpc.so.6.0 00:05:26.957 SYMLINK libspdk_rpc.so 00:05:26.957 SYMLINK libspdk_env_dpdk.so 00:05:27.215 CC lib/trace/trace_flags.o 00:05:27.215 CC lib/trace/trace.o 00:05:27.215 CC lib/trace/trace_rpc.o 00:05:27.215 CC lib/notify/notify.o 00:05:27.215 CC lib/notify/notify_rpc.o 00:05:27.215 CC lib/keyring/keyring_rpc.o 00:05:27.215 CC lib/keyring/keyring.o 00:05:27.472 LIB libspdk_notify.a 00:05:27.472 SO libspdk_notify.so.6.0 00:05:27.472 LIB libspdk_keyring.a 00:05:27.472 LIB libspdk_trace.a 00:05:27.472 SYMLINK libspdk_notify.so 00:05:27.472 SO libspdk_keyring.so.2.0 00:05:27.472 SO libspdk_trace.so.11.0 00:05:27.472 SYMLINK libspdk_keyring.so 00:05:27.472 SYMLINK libspdk_trace.so 00:05:27.729 CC lib/sock/sock_rpc.o 00:05:27.729 CC lib/thread/thread.o 00:05:27.729 CC lib/sock/sock.o 00:05:27.729 CC lib/thread/iobuf.o 00:05:27.989 LIB libspdk_sock.a 00:05:27.989 SO libspdk_sock.so.10.0 00:05:28.247 SYMLINK libspdk_sock.so 00:05:28.247 CC lib/nvme/nvme_ctrlr_cmd.o 00:05:28.247 CC lib/nvme/nvme_ctrlr.o 00:05:28.247 CC lib/nvme/nvme_fabric.o 00:05:28.247 CC lib/nvme/nvme_ns_cmd.o 00:05:28.247 CC lib/nvme/nvme_qpair.o 00:05:28.247 CC lib/nvme/nvme_ns.o 00:05:28.247 CC lib/nvme/nvme_pcie_common.o 00:05:28.247 CC lib/nvme/nvme_pcie.o 00:05:28.247 CC lib/nvme/nvme.o 00:05:29.180 CC lib/nvme/nvme_quirks.o 00:05:29.180 CC lib/nvme/nvme_transport.o 00:05:29.180 CC lib/nvme/nvme_discovery.o 00:05:29.180 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:05:29.180 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:05:29.180 CC lib/nvme/nvme_tcp.o 00:05:29.180 CC lib/nvme/nvme_opal.o 00:05:29.180 CC lib/nvme/nvme_io_msg.o 00:05:29.180 LIB libspdk_thread.a 00:05:29.180 SO libspdk_thread.so.11.0 00:05:29.437 SYMLINK libspdk_thread.so 00:05:29.437 CC lib/nvme/nvme_poll_group.o 00:05:29.437 CC lib/nvme/nvme_zns.o 00:05:29.437 CC lib/nvme/nvme_stubs.o 00:05:29.437 CC lib/nvme/nvme_auth.o 00:05:29.437 CC lib/nvme/nvme_cuse.o 00:05:29.697 CC lib/nvme/nvme_rdma.o 00:05:29.697 CC lib/accel/accel.o 00:05:29.697 CC lib/blob/blobstore.o 00:05:29.960 CC lib/blob/request.o 00:05:29.960 CC lib/blob/zeroes.o 00:05:29.960 CC lib/init/json_config.o 00:05:29.960 CC lib/init/subsystem.o 00:05:30.219 CC lib/blob/blob_bs_dev.o 00:05:30.219 CC lib/init/subsystem_rpc.o 00:05:30.219 CC lib/init/rpc.o 00:05:30.529 CC lib/accel/accel_rpc.o 00:05:30.529 LIB libspdk_init.a 00:05:30.529 CC lib/accel/accel_sw.o 00:05:30.529 SO libspdk_init.so.6.0 00:05:30.529 CC lib/fsdev/fsdev.o 00:05:30.529 CC lib/virtio/virtio.o 00:05:30.529 CC lib/virtio/virtio_vhost_user.o 00:05:30.529 SYMLINK libspdk_init.so 00:05:30.529 CC lib/virtio/virtio_vfio_user.o 00:05:30.529 CC lib/virtio/virtio_pci.o 00:05:30.787 CC lib/fsdev/fsdev_io.o 00:05:30.787 CC lib/fsdev/fsdev_rpc.o 00:05:30.787 CC lib/event/app.o 00:05:30.787 CC lib/event/reactor.o 00:05:30.787 CC lib/event/log_rpc.o 00:05:30.787 LIB libspdk_virtio.a 00:05:31.043 LIB libspdk_accel.a 00:05:31.043 CC lib/event/app_rpc.o 00:05:31.043 SO libspdk_virtio.so.7.0 00:05:31.043 SO libspdk_accel.so.16.0 00:05:31.043 CC lib/event/scheduler_static.o 00:05:31.043 SYMLINK libspdk_virtio.so 00:05:31.043 LIB libspdk_nvme.a 00:05:31.043 SYMLINK libspdk_accel.so 00:05:31.043 LIB libspdk_fsdev.a 00:05:31.043 SO libspdk_nvme.so.15.0 00:05:31.300 SO libspdk_fsdev.so.2.0 00:05:31.300 CC lib/bdev/bdev.o 00:05:31.300 CC lib/bdev/bdev_rpc.o 00:05:31.300 CC lib/bdev/bdev_zone.o 00:05:31.300 CC lib/bdev/part.o 00:05:31.300 CC lib/bdev/scsi_nvme.o 00:05:31.300 SYMLINK libspdk_fsdev.so 00:05:31.300 LIB libspdk_event.a 00:05:31.300 SO libspdk_event.so.14.0 00:05:31.300 SYMLINK libspdk_event.so 00:05:31.300 SYMLINK libspdk_nvme.so 00:05:31.300 CC lib/fuse_dispatcher/fuse_dispatcher.o 00:05:32.234 LIB libspdk_fuse_dispatcher.a 00:05:32.234 SO libspdk_fuse_dispatcher.so.1.0 00:05:32.234 SYMLINK libspdk_fuse_dispatcher.so 00:05:33.167 LIB libspdk_blob.a 00:05:33.167 SO libspdk_blob.so.11.0 00:05:33.167 SYMLINK libspdk_blob.so 00:05:33.425 CC lib/blobfs/tree.o 00:05:33.425 CC lib/blobfs/blobfs.o 00:05:33.425 CC lib/lvol/lvol.o 00:05:33.997 LIB libspdk_bdev.a 00:05:33.997 SO libspdk_bdev.so.17.0 00:05:33.997 SYMLINK libspdk_bdev.so 00:05:34.254 CC lib/nvmf/ctrlr.o 00:05:34.254 CC lib/nvmf/ctrlr_discovery.o 00:05:34.254 CC lib/nvmf/ctrlr_bdev.o 00:05:34.254 CC lib/ublk/ublk.o 00:05:34.254 CC lib/ublk/ublk_rpc.o 00:05:34.254 CC lib/nbd/nbd.o 00:05:34.254 CC lib/scsi/dev.o 00:05:34.254 CC lib/ftl/ftl_core.o 00:05:34.254 LIB libspdk_blobfs.a 00:05:34.254 SO libspdk_blobfs.so.10.0 00:05:34.254 SYMLINK libspdk_blobfs.so 00:05:34.254 CC lib/ftl/ftl_init.o 00:05:34.512 LIB libspdk_lvol.a 00:05:34.512 CC lib/ftl/ftl_layout.o 00:05:34.512 SO libspdk_lvol.so.10.0 00:05:34.512 CC lib/scsi/lun.o 00:05:34.512 SYMLINK libspdk_lvol.so 00:05:34.512 CC lib/scsi/port.o 00:05:34.512 CC lib/scsi/scsi.o 00:05:34.512 CC lib/nvmf/subsystem.o 00:05:34.512 CC lib/nvmf/nvmf.o 00:05:34.512 CC lib/nbd/nbd_rpc.o 00:05:34.771 CC lib/scsi/scsi_bdev.o 00:05:34.771 CC lib/scsi/scsi_pr.o 00:05:34.771 CC lib/ftl/ftl_debug.o 00:05:34.771 CC lib/ftl/ftl_io.o 00:05:34.771 LIB libspdk_nbd.a 00:05:34.771 SO libspdk_nbd.so.7.0 00:05:34.771 CC lib/nvmf/nvmf_rpc.o 00:05:35.030 LIB libspdk_ublk.a 00:05:35.030 SYMLINK libspdk_nbd.so 00:05:35.030 CC lib/nvmf/transport.o 00:05:35.030 CC lib/nvmf/tcp.o 00:05:35.030 SO libspdk_ublk.so.3.0 00:05:35.030 CC lib/ftl/ftl_sb.o 00:05:35.030 CC lib/ftl/ftl_l2p.o 00:05:35.030 SYMLINK libspdk_ublk.so 00:05:35.030 CC lib/ftl/ftl_l2p_flat.o 00:05:35.030 CC lib/scsi/scsi_rpc.o 00:05:35.030 CC lib/scsi/task.o 00:05:35.287 CC lib/ftl/ftl_nv_cache.o 00:05:35.287 CC lib/ftl/ftl_band.o 00:05:35.287 CC lib/ftl/ftl_band_ops.o 00:05:35.287 LIB libspdk_scsi.a 00:05:35.287 SO libspdk_scsi.so.9.0 00:05:35.545 CC lib/ftl/ftl_writer.o 00:05:35.545 SYMLINK libspdk_scsi.so 00:05:35.545 CC lib/ftl/ftl_rq.o 00:05:35.545 CC lib/nvmf/stubs.o 00:05:35.545 CC lib/nvmf/mdns_server.o 00:05:35.545 CC lib/nvmf/rdma.o 00:05:35.545 CC lib/ftl/ftl_reloc.o 00:05:35.545 CC lib/iscsi/conn.o 00:05:35.545 CC lib/iscsi/init_grp.o 00:05:35.804 CC lib/iscsi/iscsi.o 00:05:35.804 CC lib/ftl/ftl_l2p_cache.o 00:05:35.804 CC lib/ftl/ftl_p2l.o 00:05:36.063 CC lib/nvmf/auth.o 00:05:36.063 CC lib/ftl/ftl_p2l_log.o 00:05:36.063 CC lib/vhost/vhost.o 00:05:36.322 CC lib/vhost/vhost_rpc.o 00:05:36.322 CC lib/vhost/vhost_scsi.o 00:05:36.322 CC lib/vhost/vhost_blk.o 00:05:36.322 CC lib/vhost/rte_vhost_user.o 00:05:36.581 CC lib/ftl/mngt/ftl_mngt.o 00:05:36.581 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:05:36.581 CC lib/iscsi/param.o 00:05:36.842 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:05:36.842 CC lib/iscsi/portal_grp.o 00:05:36.842 CC lib/ftl/mngt/ftl_mngt_startup.o 00:05:36.842 CC lib/ftl/mngt/ftl_mngt_md.o 00:05:36.842 CC lib/ftl/mngt/ftl_mngt_misc.o 00:05:36.843 CC lib/iscsi/tgt_node.o 00:05:37.102 CC lib/iscsi/iscsi_subsystem.o 00:05:37.102 CC lib/iscsi/iscsi_rpc.o 00:05:37.102 CC lib/iscsi/task.o 00:05:37.102 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:05:37.102 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:05:37.102 CC lib/ftl/mngt/ftl_mngt_band.o 00:05:37.102 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:05:37.360 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:05:37.360 LIB libspdk_vhost.a 00:05:37.360 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:05:37.360 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:05:37.360 SO libspdk_vhost.so.8.0 00:05:37.360 CC lib/ftl/utils/ftl_conf.o 00:05:37.360 CC lib/ftl/utils/ftl_md.o 00:05:37.360 CC lib/ftl/utils/ftl_mempool.o 00:05:37.360 SYMLINK libspdk_vhost.so 00:05:37.360 CC lib/ftl/utils/ftl_bitmap.o 00:05:37.360 CC lib/ftl/utils/ftl_property.o 00:05:37.360 LIB libspdk_iscsi.a 00:05:37.360 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:05:37.360 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:05:37.618 SO libspdk_iscsi.so.8.0 00:05:37.618 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:05:37.618 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:05:37.618 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:05:37.618 SYMLINK libspdk_iscsi.so 00:05:37.618 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:05:37.618 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:05:37.618 CC lib/ftl/upgrade/ftl_sb_v3.o 00:05:37.618 CC lib/ftl/upgrade/ftl_sb_v5.o 00:05:37.618 CC lib/ftl/nvc/ftl_nvc_dev.o 00:05:37.878 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:05:37.878 CC lib/ftl/nvc/ftl_nvc_bdev_non_vss.o 00:05:37.878 CC lib/ftl/nvc/ftl_nvc_bdev_common.o 00:05:37.878 CC lib/ftl/base/ftl_base_dev.o 00:05:37.878 CC lib/ftl/base/ftl_base_bdev.o 00:05:37.878 CC lib/ftl/ftl_trace.o 00:05:37.878 LIB libspdk_nvmf.a 00:05:37.878 LIB libspdk_ftl.a 00:05:37.878 SO libspdk_nvmf.so.20.0 00:05:38.136 SYMLINK libspdk_nvmf.so 00:05:38.136 SO libspdk_ftl.so.9.0 00:05:38.394 SYMLINK libspdk_ftl.so 00:05:38.652 CC module/env_dpdk/env_dpdk_rpc.o 00:05:38.652 CC module/keyring/file/keyring.o 00:05:38.652 CC module/accel/error/accel_error.o 00:05:38.652 CC module/sock/posix/posix.o 00:05:38.652 CC module/accel/ioat/accel_ioat.o 00:05:38.652 CC module/keyring/linux/keyring.o 00:05:38.652 CC module/blob/bdev/blob_bdev.o 00:05:38.652 CC module/scheduler/dynamic/scheduler_dynamic.o 00:05:38.652 CC module/fsdev/aio/fsdev_aio.o 00:05:38.652 CC module/accel/dsa/accel_dsa.o 00:05:38.910 LIB libspdk_env_dpdk_rpc.a 00:05:38.910 SO libspdk_env_dpdk_rpc.so.6.0 00:05:38.910 CC module/keyring/file/keyring_rpc.o 00:05:38.910 CC module/keyring/linux/keyring_rpc.o 00:05:38.910 SYMLINK libspdk_env_dpdk_rpc.so 00:05:38.910 CC module/accel/ioat/accel_ioat_rpc.o 00:05:38.910 LIB libspdk_scheduler_dynamic.a 00:05:38.910 SO libspdk_scheduler_dynamic.so.4.0 00:05:38.910 LIB libspdk_keyring_file.a 00:05:38.910 CC module/accel/error/accel_error_rpc.o 00:05:38.910 LIB libspdk_keyring_linux.a 00:05:38.910 SO libspdk_keyring_file.so.2.0 00:05:38.910 SYMLINK libspdk_scheduler_dynamic.so 00:05:38.910 SO libspdk_keyring_linux.so.1.0 00:05:38.910 LIB libspdk_accel_ioat.a 00:05:38.910 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:05:38.910 SO libspdk_accel_ioat.so.6.0 00:05:38.910 LIB libspdk_blob_bdev.a 00:05:38.910 SYMLINK libspdk_keyring_linux.so 00:05:38.910 SYMLINK libspdk_keyring_file.so 00:05:38.910 SO libspdk_blob_bdev.so.11.0 00:05:38.910 CC module/accel/dsa/accel_dsa_rpc.o 00:05:38.910 CC module/fsdev/aio/fsdev_aio_rpc.o 00:05:39.169 SYMLINK libspdk_accel_ioat.so 00:05:39.169 CC module/fsdev/aio/linux_aio_mgr.o 00:05:39.169 LIB libspdk_accel_error.a 00:05:39.169 SYMLINK libspdk_blob_bdev.so 00:05:39.169 SO libspdk_accel_error.so.2.0 00:05:39.169 CC module/scheduler/gscheduler/gscheduler.o 00:05:39.169 LIB libspdk_scheduler_dpdk_governor.a 00:05:39.169 SO libspdk_scheduler_dpdk_governor.so.4.0 00:05:39.169 CC module/accel/iaa/accel_iaa.o 00:05:39.169 LIB libspdk_accel_dsa.a 00:05:39.169 SYMLINK libspdk_accel_error.so 00:05:39.169 SYMLINK libspdk_scheduler_dpdk_governor.so 00:05:39.169 SO libspdk_accel_dsa.so.5.0 00:05:39.169 CC module/accel/iaa/accel_iaa_rpc.o 00:05:39.169 SYMLINK libspdk_accel_dsa.so 00:05:39.169 LIB libspdk_scheduler_gscheduler.a 00:05:39.169 SO libspdk_scheduler_gscheduler.so.4.0 00:05:39.429 LIB libspdk_fsdev_aio.a 00:05:39.429 SYMLINK libspdk_scheduler_gscheduler.so 00:05:39.429 CC module/bdev/delay/vbdev_delay.o 00:05:39.429 CC module/bdev/error/vbdev_error.o 00:05:39.429 LIB libspdk_accel_iaa.a 00:05:39.429 SO libspdk_fsdev_aio.so.1.0 00:05:39.429 CC module/bdev/gpt/gpt.o 00:05:39.429 SO libspdk_accel_iaa.so.3.0 00:05:39.429 CC module/blobfs/bdev/blobfs_bdev.o 00:05:39.429 SYMLINK libspdk_fsdev_aio.so 00:05:39.429 CC module/bdev/lvol/vbdev_lvol.o 00:05:39.429 SYMLINK libspdk_accel_iaa.so 00:05:39.429 CC module/bdev/malloc/bdev_malloc.o 00:05:39.429 LIB libspdk_sock_posix.a 00:05:39.429 CC module/bdev/null/bdev_null.o 00:05:39.429 SO libspdk_sock_posix.so.6.0 00:05:39.429 CC module/bdev/gpt/vbdev_gpt.o 00:05:39.429 CC module/bdev/nvme/bdev_nvme.o 00:05:39.429 CC module/bdev/passthru/vbdev_passthru.o 00:05:39.429 SYMLINK libspdk_sock_posix.so 00:05:39.693 CC module/bdev/error/vbdev_error_rpc.o 00:05:39.693 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:05:39.693 CC module/bdev/delay/vbdev_delay_rpc.o 00:05:39.693 CC module/bdev/null/bdev_null_rpc.o 00:05:39.693 CC module/bdev/raid/bdev_raid.o 00:05:39.693 LIB libspdk_bdev_error.a 00:05:39.693 LIB libspdk_blobfs_bdev.a 00:05:39.693 LIB libspdk_bdev_delay.a 00:05:39.693 SO libspdk_blobfs_bdev.so.6.0 00:05:39.693 LIB libspdk_bdev_gpt.a 00:05:39.693 SO libspdk_bdev_delay.so.6.0 00:05:39.693 SO libspdk_bdev_error.so.6.0 00:05:39.693 CC module/bdev/malloc/bdev_malloc_rpc.o 00:05:39.693 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:05:39.693 SO libspdk_bdev_gpt.so.6.0 00:05:39.693 SYMLINK libspdk_bdev_delay.so 00:05:39.693 SYMLINK libspdk_blobfs_bdev.so 00:05:39.693 SYMLINK libspdk_bdev_error.so 00:05:39.693 CC module/bdev/nvme/bdev_nvme_rpc.o 00:05:39.693 CC module/bdev/nvme/nvme_rpc.o 00:05:39.693 CC module/bdev/nvme/bdev_mdns_client.o 00:05:39.952 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:05:39.952 LIB libspdk_bdev_null.a 00:05:39.952 SYMLINK libspdk_bdev_gpt.so 00:05:39.952 SO libspdk_bdev_null.so.6.0 00:05:39.952 LIB libspdk_bdev_malloc.a 00:05:39.952 LIB libspdk_bdev_passthru.a 00:05:39.952 SO libspdk_bdev_malloc.so.6.0 00:05:39.952 SYMLINK libspdk_bdev_null.so 00:05:39.952 SO libspdk_bdev_passthru.so.6.0 00:05:39.952 SYMLINK libspdk_bdev_passthru.so 00:05:39.952 SYMLINK libspdk_bdev_malloc.so 00:05:39.953 CC module/bdev/raid/bdev_raid_rpc.o 00:05:39.953 CC module/bdev/raid/bdev_raid_sb.o 00:05:39.953 CC module/bdev/split/vbdev_split.o 00:05:39.953 CC module/bdev/split/vbdev_split_rpc.o 00:05:40.213 CC module/bdev/zone_block/vbdev_zone_block.o 00:05:40.213 CC module/bdev/aio/bdev_aio.o 00:05:40.213 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:05:40.213 CC module/bdev/raid/raid0.o 00:05:40.213 LIB libspdk_bdev_lvol.a 00:05:40.213 LIB libspdk_bdev_split.a 00:05:40.213 SO libspdk_bdev_lvol.so.6.0 00:05:40.213 SO libspdk_bdev_split.so.6.0 00:05:40.213 CC module/bdev/nvme/vbdev_opal.o 00:05:40.213 SYMLINK libspdk_bdev_split.so 00:05:40.213 SYMLINK libspdk_bdev_lvol.so 00:05:40.213 CC module/bdev/raid/raid1.o 00:05:40.472 CC module/bdev/ftl/bdev_ftl.o 00:05:40.472 LIB libspdk_bdev_zone_block.a 00:05:40.472 SO libspdk_bdev_zone_block.so.6.0 00:05:40.473 CC module/bdev/iscsi/bdev_iscsi.o 00:05:40.473 SYMLINK libspdk_bdev_zone_block.so 00:05:40.473 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:05:40.473 CC module/bdev/raid/concat.o 00:05:40.473 CC module/bdev/ftl/bdev_ftl_rpc.o 00:05:40.473 CC module/bdev/aio/bdev_aio_rpc.o 00:05:40.473 CC module/bdev/raid/raid5f.o 00:05:40.473 CC module/bdev/nvme/vbdev_opal_rpc.o 00:05:40.473 CC module/bdev/virtio/bdev_virtio_scsi.o 00:05:40.730 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:05:40.730 LIB libspdk_bdev_aio.a 00:05:40.730 SO libspdk_bdev_aio.so.6.0 00:05:40.730 LIB libspdk_bdev_ftl.a 00:05:40.730 CC module/bdev/virtio/bdev_virtio_blk.o 00:05:40.730 SO libspdk_bdev_ftl.so.6.0 00:05:40.730 SYMLINK libspdk_bdev_aio.so 00:05:40.730 CC module/bdev/virtio/bdev_virtio_rpc.o 00:05:40.730 SYMLINK libspdk_bdev_ftl.so 00:05:40.730 LIB libspdk_bdev_iscsi.a 00:05:40.730 SO libspdk_bdev_iscsi.so.6.0 00:05:40.988 SYMLINK libspdk_bdev_iscsi.so 00:05:40.988 LIB libspdk_bdev_raid.a 00:05:40.988 SO libspdk_bdev_raid.so.6.0 00:05:40.988 SYMLINK libspdk_bdev_raid.so 00:05:40.988 LIB libspdk_bdev_virtio.a 00:05:41.245 SO libspdk_bdev_virtio.so.6.0 00:05:41.245 SYMLINK libspdk_bdev_virtio.so 00:05:41.808 LIB libspdk_bdev_nvme.a 00:05:41.808 SO libspdk_bdev_nvme.so.7.1 00:05:42.065 SYMLINK libspdk_bdev_nvme.so 00:05:42.322 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:05:42.322 CC module/event/subsystems/vmd/vmd.o 00:05:42.322 CC module/event/subsystems/vmd/vmd_rpc.o 00:05:42.322 CC module/event/subsystems/keyring/keyring.o 00:05:42.322 CC module/event/subsystems/scheduler/scheduler.o 00:05:42.322 CC module/event/subsystems/iobuf/iobuf.o 00:05:42.322 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:05:42.322 CC module/event/subsystems/fsdev/fsdev.o 00:05:42.322 CC module/event/subsystems/sock/sock.o 00:05:42.322 LIB libspdk_event_vhost_blk.a 00:05:42.322 LIB libspdk_event_keyring.a 00:05:42.322 LIB libspdk_event_fsdev.a 00:05:42.322 LIB libspdk_event_vmd.a 00:05:42.322 LIB libspdk_event_scheduler.a 00:05:42.579 SO libspdk_event_vhost_blk.so.3.0 00:05:42.579 SO libspdk_event_fsdev.so.1.0 00:05:42.579 LIB libspdk_event_sock.a 00:05:42.579 SO libspdk_event_keyring.so.1.0 00:05:42.579 LIB libspdk_event_iobuf.a 00:05:42.579 SO libspdk_event_vmd.so.6.0 00:05:42.579 SO libspdk_event_sock.so.5.0 00:05:42.579 SO libspdk_event_scheduler.so.4.0 00:05:42.579 SO libspdk_event_iobuf.so.3.0 00:05:42.579 SYMLINK libspdk_event_fsdev.so 00:05:42.579 SYMLINK libspdk_event_vhost_blk.so 00:05:42.579 SYMLINK libspdk_event_keyring.so 00:05:42.579 SYMLINK libspdk_event_sock.so 00:05:42.579 SYMLINK libspdk_event_vmd.so 00:05:42.579 SYMLINK libspdk_event_scheduler.so 00:05:42.579 SYMLINK libspdk_event_iobuf.so 00:05:42.835 CC module/event/subsystems/accel/accel.o 00:05:42.835 LIB libspdk_event_accel.a 00:05:42.835 SO libspdk_event_accel.so.6.0 00:05:42.835 SYMLINK libspdk_event_accel.so 00:05:43.091 CC module/event/subsystems/bdev/bdev.o 00:05:43.347 LIB libspdk_event_bdev.a 00:05:43.347 SO libspdk_event_bdev.so.6.0 00:05:43.347 SYMLINK libspdk_event_bdev.so 00:05:43.604 CC module/event/subsystems/scsi/scsi.o 00:05:43.604 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:05:43.604 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:05:43.604 CC module/event/subsystems/ublk/ublk.o 00:05:43.604 CC module/event/subsystems/nbd/nbd.o 00:05:43.604 LIB libspdk_event_ublk.a 00:05:43.604 LIB libspdk_event_scsi.a 00:05:43.604 SO libspdk_event_ublk.so.3.0 00:05:43.604 LIB libspdk_event_nbd.a 00:05:43.604 SO libspdk_event_scsi.so.6.0 00:05:43.604 SO libspdk_event_nbd.so.6.0 00:05:43.605 LIB libspdk_event_nvmf.a 00:05:43.605 SYMLINK libspdk_event_ublk.so 00:05:43.605 SYMLINK libspdk_event_scsi.so 00:05:43.605 SO libspdk_event_nvmf.so.6.0 00:05:43.605 SYMLINK libspdk_event_nbd.so 00:05:43.862 SYMLINK libspdk_event_nvmf.so 00:05:43.862 CC module/event/subsystems/iscsi/iscsi.o 00:05:43.862 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:05:44.120 LIB libspdk_event_iscsi.a 00:05:44.120 LIB libspdk_event_vhost_scsi.a 00:05:44.120 SO libspdk_event_iscsi.so.6.0 00:05:44.120 SO libspdk_event_vhost_scsi.so.3.0 00:05:44.120 SYMLINK libspdk_event_iscsi.so 00:05:44.120 SYMLINK libspdk_event_vhost_scsi.so 00:05:44.120 SO libspdk.so.6.0 00:05:44.120 SYMLINK libspdk.so 00:05:44.378 CXX app/trace/trace.o 00:05:44.378 CC app/spdk_nvme_perf/perf.o 00:05:44.378 CC app/trace_record/trace_record.o 00:05:44.378 CC app/spdk_lspci/spdk_lspci.o 00:05:44.378 CC app/spdk_nvme_identify/identify.o 00:05:44.378 CC app/nvmf_tgt/nvmf_main.o 00:05:44.378 CC app/iscsi_tgt/iscsi_tgt.o 00:05:44.378 CC app/spdk_tgt/spdk_tgt.o 00:05:44.378 CC examples/util/zipf/zipf.o 00:05:44.378 CC test/thread/poller_perf/poller_perf.o 00:05:44.378 LINK spdk_lspci 00:05:44.636 LINK nvmf_tgt 00:05:44.636 LINK iscsi_tgt 00:05:44.636 LINK zipf 00:05:44.636 LINK poller_perf 00:05:44.636 LINK spdk_tgt 00:05:44.636 CC app/spdk_nvme_discover/discovery_aer.o 00:05:44.636 LINK spdk_trace_record 00:05:44.636 CC app/spdk_top/spdk_top.o 00:05:44.636 LINK spdk_trace 00:05:44.893 CC app/spdk_dd/spdk_dd.o 00:05:44.893 CC examples/ioat/verify/verify.o 00:05:44.893 CC examples/ioat/perf/perf.o 00:05:44.893 LINK spdk_nvme_discover 00:05:44.893 CC test/dma/test_dma/test_dma.o 00:05:44.893 CC app/fio/nvme/fio_plugin.o 00:05:44.893 LINK verify 00:05:44.893 LINK ioat_perf 00:05:44.893 CC test/app/bdev_svc/bdev_svc.o 00:05:45.150 LINK spdk_nvme_identify 00:05:45.150 LINK spdk_dd 00:05:45.150 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:05:45.150 LINK bdev_svc 00:05:45.150 CC app/vhost/vhost.o 00:05:45.150 LINK spdk_nvme_perf 00:05:45.150 CC examples/vmd/lsvmd/lsvmd.o 00:05:45.408 CC examples/idxd/perf/perf.o 00:05:45.408 LINK lsvmd 00:05:45.408 CC examples/vmd/led/led.o 00:05:45.408 LINK vhost 00:05:45.408 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:05:45.408 LINK test_dma 00:05:45.408 CC examples/interrupt_tgt/interrupt_tgt.o 00:05:45.408 LINK spdk_top 00:05:45.408 LINK nvme_fuzz 00:05:45.408 LINK spdk_nvme 00:05:45.408 LINK led 00:05:45.666 TEST_HEADER include/spdk/accel.h 00:05:45.666 TEST_HEADER include/spdk/accel_module.h 00:05:45.666 TEST_HEADER include/spdk/assert.h 00:05:45.666 TEST_HEADER include/spdk/barrier.h 00:05:45.666 TEST_HEADER include/spdk/base64.h 00:05:45.667 TEST_HEADER include/spdk/bdev.h 00:05:45.667 TEST_HEADER include/spdk/bdev_module.h 00:05:45.667 TEST_HEADER include/spdk/bdev_zone.h 00:05:45.667 TEST_HEADER include/spdk/bit_array.h 00:05:45.667 TEST_HEADER include/spdk/bit_pool.h 00:05:45.667 TEST_HEADER include/spdk/blob_bdev.h 00:05:45.667 TEST_HEADER include/spdk/blobfs_bdev.h 00:05:45.667 TEST_HEADER include/spdk/blobfs.h 00:05:45.667 TEST_HEADER include/spdk/blob.h 00:05:45.667 TEST_HEADER include/spdk/conf.h 00:05:45.667 TEST_HEADER include/spdk/config.h 00:05:45.667 TEST_HEADER include/spdk/cpuset.h 00:05:45.667 TEST_HEADER include/spdk/crc16.h 00:05:45.667 TEST_HEADER include/spdk/crc32.h 00:05:45.667 TEST_HEADER include/spdk/crc64.h 00:05:45.667 LINK interrupt_tgt 00:05:45.667 TEST_HEADER include/spdk/dif.h 00:05:45.667 TEST_HEADER include/spdk/dma.h 00:05:45.667 TEST_HEADER include/spdk/endian.h 00:05:45.667 TEST_HEADER include/spdk/env_dpdk.h 00:05:45.667 TEST_HEADER include/spdk/env.h 00:05:45.667 TEST_HEADER include/spdk/event.h 00:05:45.667 TEST_HEADER include/spdk/fd_group.h 00:05:45.667 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:05:45.667 TEST_HEADER include/spdk/fd.h 00:05:45.667 LINK idxd_perf 00:05:45.667 TEST_HEADER include/spdk/file.h 00:05:45.667 TEST_HEADER include/spdk/fsdev.h 00:05:45.667 TEST_HEADER include/spdk/fsdev_module.h 00:05:45.667 TEST_HEADER include/spdk/ftl.h 00:05:45.667 TEST_HEADER include/spdk/fuse_dispatcher.h 00:05:45.667 TEST_HEADER include/spdk/gpt_spec.h 00:05:45.667 TEST_HEADER include/spdk/hexlify.h 00:05:45.667 CC app/fio/bdev/fio_plugin.o 00:05:45.667 TEST_HEADER include/spdk/histogram_data.h 00:05:45.667 TEST_HEADER include/spdk/idxd.h 00:05:45.667 TEST_HEADER include/spdk/idxd_spec.h 00:05:45.667 TEST_HEADER include/spdk/init.h 00:05:45.667 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:05:45.667 TEST_HEADER include/spdk/ioat.h 00:05:45.667 TEST_HEADER include/spdk/ioat_spec.h 00:05:45.667 TEST_HEADER include/spdk/iscsi_spec.h 00:05:45.667 TEST_HEADER include/spdk/json.h 00:05:45.667 TEST_HEADER include/spdk/jsonrpc.h 00:05:45.667 TEST_HEADER include/spdk/keyring.h 00:05:45.667 TEST_HEADER include/spdk/keyring_module.h 00:05:45.667 TEST_HEADER include/spdk/likely.h 00:05:45.667 TEST_HEADER include/spdk/log.h 00:05:45.667 TEST_HEADER include/spdk/lvol.h 00:05:45.667 TEST_HEADER include/spdk/md5.h 00:05:45.667 TEST_HEADER include/spdk/memory.h 00:05:45.667 TEST_HEADER include/spdk/mmio.h 00:05:45.667 TEST_HEADER include/spdk/nbd.h 00:05:45.667 TEST_HEADER include/spdk/net.h 00:05:45.667 TEST_HEADER include/spdk/notify.h 00:05:45.667 TEST_HEADER include/spdk/nvme.h 00:05:45.667 TEST_HEADER include/spdk/nvme_intel.h 00:05:45.667 TEST_HEADER include/spdk/nvme_ocssd.h 00:05:45.667 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:05:45.667 TEST_HEADER include/spdk/nvme_spec.h 00:05:45.667 TEST_HEADER include/spdk/nvme_zns.h 00:05:45.667 TEST_HEADER include/spdk/nvmf_cmd.h 00:05:45.667 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:05:45.667 TEST_HEADER include/spdk/nvmf.h 00:05:45.667 TEST_HEADER include/spdk/nvmf_spec.h 00:05:45.667 TEST_HEADER include/spdk/nvmf_transport.h 00:05:45.667 TEST_HEADER include/spdk/opal.h 00:05:45.667 CC examples/sock/hello_world/hello_sock.o 00:05:45.667 TEST_HEADER include/spdk/opal_spec.h 00:05:45.667 TEST_HEADER include/spdk/pci_ids.h 00:05:45.667 TEST_HEADER include/spdk/pipe.h 00:05:45.667 TEST_HEADER include/spdk/queue.h 00:05:45.667 TEST_HEADER include/spdk/reduce.h 00:05:45.667 TEST_HEADER include/spdk/rpc.h 00:05:45.667 TEST_HEADER include/spdk/scheduler.h 00:05:45.667 CC test/app/histogram_perf/histogram_perf.o 00:05:45.667 TEST_HEADER include/spdk/scsi.h 00:05:45.667 TEST_HEADER include/spdk/scsi_spec.h 00:05:45.667 CC examples/thread/thread/thread_ex.o 00:05:45.667 TEST_HEADER include/spdk/sock.h 00:05:45.667 TEST_HEADER include/spdk/stdinc.h 00:05:45.667 TEST_HEADER include/spdk/string.h 00:05:45.667 TEST_HEADER include/spdk/thread.h 00:05:45.667 TEST_HEADER include/spdk/trace.h 00:05:45.667 TEST_HEADER include/spdk/trace_parser.h 00:05:45.667 TEST_HEADER include/spdk/tree.h 00:05:45.667 TEST_HEADER include/spdk/ublk.h 00:05:45.667 TEST_HEADER include/spdk/util.h 00:05:45.667 TEST_HEADER include/spdk/uuid.h 00:05:45.667 TEST_HEADER include/spdk/version.h 00:05:45.667 TEST_HEADER include/spdk/vfio_user_pci.h 00:05:45.667 TEST_HEADER include/spdk/vfio_user_spec.h 00:05:45.667 TEST_HEADER include/spdk/vhost.h 00:05:45.667 TEST_HEADER include/spdk/vmd.h 00:05:45.667 TEST_HEADER include/spdk/xor.h 00:05:45.667 TEST_HEADER include/spdk/zipf.h 00:05:45.667 CXX test/cpp_headers/accel.o 00:05:45.925 CC test/app/jsoncat/jsoncat.o 00:05:45.925 CXX test/cpp_headers/accel_module.o 00:05:45.925 CC test/event/event_perf/event_perf.o 00:05:45.925 LINK histogram_perf 00:05:45.925 LINK jsoncat 00:05:45.925 LINK thread 00:05:45.925 LINK hello_sock 00:05:45.925 CC test/env/mem_callbacks/mem_callbacks.o 00:05:45.925 CXX test/cpp_headers/assert.o 00:05:45.925 CXX test/cpp_headers/barrier.o 00:05:45.925 LINK event_perf 00:05:46.183 LINK vhost_fuzz 00:05:46.183 LINK spdk_bdev 00:05:46.183 CC test/app/stub/stub.o 00:05:46.183 CC test/event/reactor/reactor.o 00:05:46.183 CC test/env/vtophys/vtophys.o 00:05:46.183 CXX test/cpp_headers/base64.o 00:05:46.183 CC test/nvme/aer/aer.o 00:05:46.183 CC examples/accel/perf/accel_perf.o 00:05:46.183 LINK stub 00:05:46.440 CC test/rpc_client/rpc_client_test.o 00:05:46.440 CC test/nvme/reset/reset.o 00:05:46.440 LINK vtophys 00:05:46.440 CXX test/cpp_headers/bdev.o 00:05:46.440 LINK reactor 00:05:46.440 LINK aer 00:05:46.440 LINK rpc_client_test 00:05:46.440 CC test/event/reactor_perf/reactor_perf.o 00:05:46.697 LINK reset 00:05:46.697 CXX test/cpp_headers/bdev_module.o 00:05:46.697 CC test/accel/dif/dif.o 00:05:46.697 LINK mem_callbacks 00:05:46.697 CC test/blobfs/mkfs/mkfs.o 00:05:46.697 LINK reactor_perf 00:05:46.697 CXX test/cpp_headers/bdev_zone.o 00:05:46.697 CC test/nvme/sgl/sgl.o 00:05:46.697 LINK accel_perf 00:05:46.697 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:05:46.697 CC examples/blob/hello_world/hello_blob.o 00:05:46.697 CC test/lvol/esnap/esnap.o 00:05:46.954 LINK mkfs 00:05:46.954 CC test/event/app_repeat/app_repeat.o 00:05:46.954 LINK env_dpdk_post_init 00:05:46.954 CXX test/cpp_headers/bit_array.o 00:05:46.954 LINK sgl 00:05:46.955 CC test/env/memory/memory_ut.o 00:05:46.955 LINK hello_blob 00:05:46.955 CC test/env/pci/pci_ut.o 00:05:46.955 LINK app_repeat 00:05:46.955 CXX test/cpp_headers/bit_pool.o 00:05:47.212 CC test/nvme/e2edp/nvme_dp.o 00:05:47.212 CC examples/blob/cli/blobcli.o 00:05:47.212 LINK iscsi_fuzz 00:05:47.212 CXX test/cpp_headers/blob_bdev.o 00:05:47.212 CC test/event/scheduler/scheduler.o 00:05:47.212 LINK dif 00:05:47.212 CC examples/nvme/hello_world/hello_world.o 00:05:47.212 CXX test/cpp_headers/blobfs_bdev.o 00:05:47.212 CXX test/cpp_headers/blobfs.o 00:05:47.470 LINK nvme_dp 00:05:47.470 LINK scheduler 00:05:47.470 CXX test/cpp_headers/blob.o 00:05:47.470 LINK pci_ut 00:05:47.470 LINK blobcli 00:05:47.470 LINK hello_world 00:05:47.470 CC examples/fsdev/hello_world/hello_fsdev.o 00:05:47.730 CC test/nvme/overhead/overhead.o 00:05:47.730 CC test/nvme/err_injection/err_injection.o 00:05:47.730 CXX test/cpp_headers/conf.o 00:05:47.730 CC examples/bdev/hello_world/hello_bdev.o 00:05:47.730 CC test/nvme/startup/startup.o 00:05:47.730 CC examples/nvme/reconnect/reconnect.o 00:05:47.730 CC test/nvme/reserve/reserve.o 00:05:47.730 CXX test/cpp_headers/config.o 00:05:47.730 CXX test/cpp_headers/cpuset.o 00:05:47.730 LINK hello_fsdev 00:05:47.730 LINK err_injection 00:05:47.989 LINK hello_bdev 00:05:47.989 LINK overhead 00:05:47.989 LINK startup 00:05:47.989 LINK reserve 00:05:47.989 CXX test/cpp_headers/crc16.o 00:05:47.989 CXX test/cpp_headers/crc32.o 00:05:47.989 CC test/nvme/simple_copy/simple_copy.o 00:05:47.989 CC examples/bdev/bdevperf/bdevperf.o 00:05:47.989 LINK reconnect 00:05:47.989 LINK memory_ut 00:05:47.989 CC examples/nvme/arbitration/arbitration.o 00:05:47.989 CC examples/nvme/nvme_manage/nvme_manage.o 00:05:48.295 CXX test/cpp_headers/crc64.o 00:05:48.295 CC test/nvme/connect_stress/connect_stress.o 00:05:48.295 CC test/bdev/bdevio/bdevio.o 00:05:48.295 LINK simple_copy 00:05:48.295 CXX test/cpp_headers/dif.o 00:05:48.295 CC test/nvme/boot_partition/boot_partition.o 00:05:48.295 CC test/nvme/compliance/nvme_compliance.o 00:05:48.295 LINK connect_stress 00:05:48.552 LINK arbitration 00:05:48.552 CXX test/cpp_headers/dma.o 00:05:48.552 LINK boot_partition 00:05:48.552 CC examples/nvme/hotplug/hotplug.o 00:05:48.552 CC test/nvme/fused_ordering/fused_ordering.o 00:05:48.552 CXX test/cpp_headers/endian.o 00:05:48.552 CXX test/cpp_headers/env_dpdk.o 00:05:48.552 CC examples/nvme/cmb_copy/cmb_copy.o 00:05:48.810 CXX test/cpp_headers/env.o 00:05:48.810 CXX test/cpp_headers/event.o 00:05:48.810 LINK bdevio 00:05:48.810 LINK hotplug 00:05:48.810 LINK fused_ordering 00:05:48.810 LINK nvme_compliance 00:05:48.810 LINK nvme_manage 00:05:48.810 LINK cmb_copy 00:05:48.810 CXX test/cpp_headers/fd_group.o 00:05:48.810 CXX test/cpp_headers/fd.o 00:05:48.810 CXX test/cpp_headers/file.o 00:05:48.810 CC examples/nvme/abort/abort.o 00:05:48.810 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:05:49.068 CC test/nvme/doorbell_aers/doorbell_aers.o 00:05:49.068 CC test/nvme/fdp/fdp.o 00:05:49.068 CC test/nvme/cuse/cuse.o 00:05:49.068 CXX test/cpp_headers/fsdev.o 00:05:49.068 CXX test/cpp_headers/fsdev_module.o 00:05:49.068 LINK bdevperf 00:05:49.068 CXX test/cpp_headers/ftl.o 00:05:49.068 LINK pmr_persistence 00:05:49.068 CXX test/cpp_headers/fuse_dispatcher.o 00:05:49.068 LINK doorbell_aers 00:05:49.327 CXX test/cpp_headers/gpt_spec.o 00:05:49.327 CXX test/cpp_headers/hexlify.o 00:05:49.327 CXX test/cpp_headers/histogram_data.o 00:05:49.327 CXX test/cpp_headers/idxd.o 00:05:49.327 CXX test/cpp_headers/idxd_spec.o 00:05:49.327 LINK fdp 00:05:49.327 CXX test/cpp_headers/init.o 00:05:49.327 LINK abort 00:05:49.327 CXX test/cpp_headers/ioat.o 00:05:49.327 CXX test/cpp_headers/ioat_spec.o 00:05:49.327 CXX test/cpp_headers/iscsi_spec.o 00:05:49.327 CXX test/cpp_headers/json.o 00:05:49.327 CXX test/cpp_headers/jsonrpc.o 00:05:49.327 CXX test/cpp_headers/keyring.o 00:05:49.327 CXX test/cpp_headers/keyring_module.o 00:05:49.585 CXX test/cpp_headers/likely.o 00:05:49.585 CXX test/cpp_headers/log.o 00:05:49.585 CXX test/cpp_headers/lvol.o 00:05:49.585 CXX test/cpp_headers/md5.o 00:05:49.585 CXX test/cpp_headers/memory.o 00:05:49.585 CXX test/cpp_headers/mmio.o 00:05:49.585 CXX test/cpp_headers/nbd.o 00:05:49.585 CXX test/cpp_headers/net.o 00:05:49.585 CXX test/cpp_headers/notify.o 00:05:49.585 CC examples/nvmf/nvmf/nvmf.o 00:05:49.585 CXX test/cpp_headers/nvme.o 00:05:49.585 CXX test/cpp_headers/nvme_intel.o 00:05:49.585 CXX test/cpp_headers/nvme_ocssd.o 00:05:49.585 CXX test/cpp_headers/nvme_ocssd_spec.o 00:05:49.843 CXX test/cpp_headers/nvme_spec.o 00:05:49.843 CXX test/cpp_headers/nvme_zns.o 00:05:49.843 CXX test/cpp_headers/nvmf_cmd.o 00:05:49.843 CXX test/cpp_headers/nvmf_fc_spec.o 00:05:49.843 CXX test/cpp_headers/nvmf.o 00:05:49.843 CXX test/cpp_headers/nvmf_spec.o 00:05:49.843 CXX test/cpp_headers/nvmf_transport.o 00:05:49.843 CXX test/cpp_headers/opal.o 00:05:49.843 CXX test/cpp_headers/opal_spec.o 00:05:49.843 LINK nvmf 00:05:49.843 CXX test/cpp_headers/pci_ids.o 00:05:49.843 CXX test/cpp_headers/pipe.o 00:05:50.101 CXX test/cpp_headers/queue.o 00:05:50.101 CXX test/cpp_headers/reduce.o 00:05:50.101 CXX test/cpp_headers/rpc.o 00:05:50.101 CXX test/cpp_headers/scheduler.o 00:05:50.101 CXX test/cpp_headers/scsi_spec.o 00:05:50.101 CXX test/cpp_headers/scsi.o 00:05:50.101 CXX test/cpp_headers/sock.o 00:05:50.101 CXX test/cpp_headers/stdinc.o 00:05:50.101 CXX test/cpp_headers/string.o 00:05:50.101 CXX test/cpp_headers/thread.o 00:05:50.101 CXX test/cpp_headers/trace.o 00:05:50.101 CXX test/cpp_headers/trace_parser.o 00:05:50.101 CXX test/cpp_headers/tree.o 00:05:50.101 CXX test/cpp_headers/ublk.o 00:05:50.101 CXX test/cpp_headers/util.o 00:05:50.101 CXX test/cpp_headers/uuid.o 00:05:50.101 LINK cuse 00:05:50.101 CXX test/cpp_headers/version.o 00:05:50.359 CXX test/cpp_headers/vfio_user_pci.o 00:05:50.359 CXX test/cpp_headers/vfio_user_spec.o 00:05:50.359 CXX test/cpp_headers/vhost.o 00:05:50.359 CXX test/cpp_headers/vmd.o 00:05:50.359 CXX test/cpp_headers/xor.o 00:05:50.359 CXX test/cpp_headers/zipf.o 00:05:52.890 LINK esnap 00:05:52.890 00:05:52.890 real 1m7.569s 00:05:52.890 user 6m15.337s 00:05:52.890 sys 1m4.818s 00:05:52.890 ************************************ 00:05:52.890 END TEST make 00:05:52.890 ************************************ 00:05:52.890 15:53:50 make -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:05:52.890 15:53:50 make -- common/autotest_common.sh@10 -- $ set +x 00:05:52.890 15:53:50 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:05:52.890 15:53:50 -- pm/common@29 -- $ signal_monitor_resources TERM 00:05:52.890 15:53:50 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:05:52.890 15:53:50 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:52.890 15:53:50 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-cpu-load.pid ]] 00:05:52.890 15:53:50 -- pm/common@44 -- $ pid=5017 00:05:52.890 15:53:50 -- pm/common@50 -- $ kill -TERM 5017 00:05:52.890 15:53:50 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:52.890 15:53:50 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-vmstat.pid ]] 00:05:52.890 15:53:50 -- pm/common@44 -- $ pid=5019 00:05:52.890 15:53:50 -- pm/common@50 -- $ kill -TERM 5019 00:05:52.890 15:53:50 -- spdk/autorun.sh@26 -- $ (( SPDK_TEST_UNITTEST == 1 || SPDK_RUN_FUNCTIONAL_TEST == 1 )) 00:05:52.890 15:53:50 -- spdk/autorun.sh@27 -- $ sudo -E /home/vagrant/spdk_repo/spdk/autotest.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:05:52.890 15:53:50 -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:52.890 15:53:50 -- common/autotest_common.sh@1693 -- # lcov --version 00:05:52.890 15:53:50 -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:52.890 15:53:51 -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:52.890 15:53:51 -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:52.890 15:53:51 -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:52.890 15:53:51 -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:52.890 15:53:51 -- scripts/common.sh@336 -- # IFS=.-: 00:05:52.890 15:53:51 -- scripts/common.sh@336 -- # read -ra ver1 00:05:52.890 15:53:51 -- scripts/common.sh@337 -- # IFS=.-: 00:05:52.890 15:53:51 -- scripts/common.sh@337 -- # read -ra ver2 00:05:52.890 15:53:51 -- scripts/common.sh@338 -- # local 'op=<' 00:05:52.890 15:53:51 -- scripts/common.sh@340 -- # ver1_l=2 00:05:52.890 15:53:51 -- scripts/common.sh@341 -- # ver2_l=1 00:05:52.890 15:53:51 -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:52.890 15:53:51 -- scripts/common.sh@344 -- # case "$op" in 00:05:52.890 15:53:51 -- scripts/common.sh@345 -- # : 1 00:05:52.890 15:53:51 -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:52.890 15:53:51 -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:52.890 15:53:51 -- scripts/common.sh@365 -- # decimal 1 00:05:52.890 15:53:51 -- scripts/common.sh@353 -- # local d=1 00:05:52.890 15:53:51 -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:52.890 15:53:51 -- scripts/common.sh@355 -- # echo 1 00:05:52.890 15:53:51 -- scripts/common.sh@365 -- # ver1[v]=1 00:05:52.890 15:53:51 -- scripts/common.sh@366 -- # decimal 2 00:05:52.890 15:53:51 -- scripts/common.sh@353 -- # local d=2 00:05:52.890 15:53:51 -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:52.890 15:53:51 -- scripts/common.sh@355 -- # echo 2 00:05:52.890 15:53:51 -- scripts/common.sh@366 -- # ver2[v]=2 00:05:52.890 15:53:51 -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:52.890 15:53:51 -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:52.890 15:53:51 -- scripts/common.sh@368 -- # return 0 00:05:52.890 15:53:51 -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:52.890 15:53:51 -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:52.890 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:52.890 --rc genhtml_branch_coverage=1 00:05:52.890 --rc genhtml_function_coverage=1 00:05:52.890 --rc genhtml_legend=1 00:05:52.890 --rc geninfo_all_blocks=1 00:05:52.890 --rc geninfo_unexecuted_blocks=1 00:05:52.890 00:05:52.890 ' 00:05:52.890 15:53:51 -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:52.890 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:52.890 --rc genhtml_branch_coverage=1 00:05:52.890 --rc genhtml_function_coverage=1 00:05:52.890 --rc genhtml_legend=1 00:05:52.890 --rc geninfo_all_blocks=1 00:05:52.890 --rc geninfo_unexecuted_blocks=1 00:05:52.890 00:05:52.890 ' 00:05:52.890 15:53:51 -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:52.890 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:52.890 --rc genhtml_branch_coverage=1 00:05:52.890 --rc genhtml_function_coverage=1 00:05:52.890 --rc genhtml_legend=1 00:05:52.890 --rc geninfo_all_blocks=1 00:05:52.890 --rc geninfo_unexecuted_blocks=1 00:05:52.890 00:05:52.890 ' 00:05:52.890 15:53:51 -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:52.890 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:52.890 --rc genhtml_branch_coverage=1 00:05:52.890 --rc genhtml_function_coverage=1 00:05:52.890 --rc genhtml_legend=1 00:05:52.890 --rc geninfo_all_blocks=1 00:05:52.890 --rc geninfo_unexecuted_blocks=1 00:05:52.890 00:05:52.890 ' 00:05:52.890 15:53:51 -- spdk/autotest.sh@25 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:05:52.890 15:53:51 -- nvmf/common.sh@7 -- # uname -s 00:05:52.890 15:53:51 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:52.890 15:53:51 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:52.890 15:53:51 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:52.890 15:53:51 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:52.890 15:53:51 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:52.890 15:53:51 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:52.890 15:53:51 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:52.890 15:53:51 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:52.890 15:53:51 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:52.890 15:53:51 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:52.890 15:53:51 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:13ef8e52-d220-4442-baed-0e105827d1f9 00:05:52.890 15:53:51 -- nvmf/common.sh@18 -- # NVME_HOSTID=13ef8e52-d220-4442-baed-0e105827d1f9 00:05:52.890 15:53:51 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:52.891 15:53:51 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:52.891 15:53:51 -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:05:52.891 15:53:51 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:52.891 15:53:51 -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:05:52.891 15:53:51 -- scripts/common.sh@15 -- # shopt -s extglob 00:05:52.891 15:53:51 -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:52.891 15:53:51 -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:52.891 15:53:51 -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:52.891 15:53:51 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:52.891 15:53:51 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:52.891 15:53:51 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:52.891 15:53:51 -- paths/export.sh@5 -- # export PATH 00:05:52.891 15:53:51 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:52.891 15:53:51 -- nvmf/common.sh@51 -- # : 0 00:05:52.891 15:53:51 -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:52.891 15:53:51 -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:52.891 15:53:51 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:52.891 15:53:51 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:52.891 15:53:51 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:52.891 15:53:51 -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:52.891 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:52.891 15:53:51 -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:52.891 15:53:51 -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:52.891 15:53:51 -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:52.891 15:53:51 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:05:52.891 15:53:51 -- spdk/autotest.sh@32 -- # uname -s 00:05:52.891 15:53:51 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:05:52.891 15:53:51 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:05:52.891 15:53:51 -- spdk/autotest.sh@34 -- # mkdir -p /home/vagrant/spdk_repo/spdk/../output/coredumps 00:05:52.891 15:53:51 -- spdk/autotest.sh@39 -- # echo '|/home/vagrant/spdk_repo/spdk/scripts/core-collector.sh %P %s %t' 00:05:52.891 15:53:51 -- spdk/autotest.sh@40 -- # echo /home/vagrant/spdk_repo/spdk/../output/coredumps 00:05:52.891 15:53:51 -- spdk/autotest.sh@44 -- # modprobe nbd 00:05:52.891 15:53:51 -- spdk/autotest.sh@46 -- # type -P udevadm 00:05:52.891 15:53:51 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:05:52.891 15:53:51 -- spdk/autotest.sh@48 -- # udevadm_pid=53697 00:05:52.891 15:53:51 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:05:52.891 15:53:51 -- pm/common@17 -- # local monitor 00:05:52.891 15:53:51 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:05:52.891 15:53:51 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:05:52.891 15:53:51 -- pm/common@25 -- # sleep 1 00:05:52.891 15:53:51 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:05:52.891 15:53:51 -- pm/common@21 -- # date +%s 00:05:52.891 15:53:51 -- pm/common@21 -- # date +%s 00:05:52.891 15:53:51 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1732118031 00:05:52.891 15:53:51 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1732118031 00:05:52.891 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1732118031_collect-cpu-load.pm.log 00:05:52.891 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1732118031_collect-vmstat.pm.log 00:05:53.865 15:53:52 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:05:53.865 15:53:52 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:05:53.865 15:53:52 -- common/autotest_common.sh@726 -- # xtrace_disable 00:05:53.865 15:53:52 -- common/autotest_common.sh@10 -- # set +x 00:05:53.865 15:53:52 -- spdk/autotest.sh@59 -- # create_test_list 00:05:53.865 15:53:52 -- common/autotest_common.sh@752 -- # xtrace_disable 00:05:53.865 15:53:52 -- common/autotest_common.sh@10 -- # set +x 00:05:54.123 15:53:52 -- spdk/autotest.sh@61 -- # dirname /home/vagrant/spdk_repo/spdk/autotest.sh 00:05:54.123 15:53:52 -- spdk/autotest.sh@61 -- # readlink -f /home/vagrant/spdk_repo/spdk 00:05:54.123 15:53:52 -- spdk/autotest.sh@61 -- # src=/home/vagrant/spdk_repo/spdk 00:05:54.123 15:53:52 -- spdk/autotest.sh@62 -- # out=/home/vagrant/spdk_repo/spdk/../output 00:05:54.123 15:53:52 -- spdk/autotest.sh@63 -- # cd /home/vagrant/spdk_repo/spdk 00:05:54.123 15:53:52 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:05:54.123 15:53:52 -- common/autotest_common.sh@1457 -- # uname 00:05:54.123 15:53:52 -- common/autotest_common.sh@1457 -- # '[' Linux = FreeBSD ']' 00:05:54.123 15:53:52 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:05:54.123 15:53:52 -- common/autotest_common.sh@1477 -- # uname 00:05:54.123 15:53:52 -- common/autotest_common.sh@1477 -- # [[ Linux = FreeBSD ]] 00:05:54.123 15:53:52 -- spdk/autotest.sh@68 -- # [[ y == y ]] 00:05:54.123 15:53:52 -- spdk/autotest.sh@70 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 --version 00:05:54.123 lcov: LCOV version 1.15 00:05:54.123 15:53:52 -- spdk/autotest.sh@72 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -i -t Baseline -d /home/vagrant/spdk_repo/spdk -o /home/vagrant/spdk_repo/spdk/../output/cov_base.info 00:06:08.999 /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:06:08.999 geninfo: WARNING: GCOV did not produce any data for /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno 00:06:23.863 15:54:20 -- spdk/autotest.sh@76 -- # timing_enter pre_cleanup 00:06:23.863 15:54:20 -- common/autotest_common.sh@726 -- # xtrace_disable 00:06:23.863 15:54:20 -- common/autotest_common.sh@10 -- # set +x 00:06:23.863 15:54:20 -- spdk/autotest.sh@78 -- # rm -f 00:06:23.863 15:54:20 -- spdk/autotest.sh@81 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:06:23.863 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:06:23.863 0000:00:11.0 (1b36 0010): Already using the nvme driver 00:06:23.863 0000:00:10.0 (1b36 0010): Already using the nvme driver 00:06:23.863 15:54:21 -- spdk/autotest.sh@83 -- # get_zoned_devs 00:06:23.863 15:54:21 -- common/autotest_common.sh@1657 -- # zoned_devs=() 00:06:23.863 15:54:21 -- common/autotest_common.sh@1657 -- # local -gA zoned_devs 00:06:23.863 15:54:21 -- common/autotest_common.sh@1658 -- # local nvme bdf 00:06:23.863 15:54:21 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:06:23.863 15:54:21 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme0n1 00:06:23.863 15:54:21 -- common/autotest_common.sh@1650 -- # local device=nvme0n1 00:06:23.863 15:54:21 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:06:23.863 15:54:21 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:06:23.863 15:54:21 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:06:23.863 15:54:21 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme1n1 00:06:23.863 15:54:21 -- common/autotest_common.sh@1650 -- # local device=nvme1n1 00:06:23.863 15:54:21 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n1/queue/zoned ]] 00:06:23.863 15:54:21 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:06:23.863 15:54:21 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:06:23.863 15:54:21 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme1n2 00:06:23.863 15:54:21 -- common/autotest_common.sh@1650 -- # local device=nvme1n2 00:06:23.863 15:54:21 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n2/queue/zoned ]] 00:06:23.863 15:54:21 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:06:23.863 15:54:21 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:06:23.863 15:54:21 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme1n3 00:06:23.863 15:54:21 -- common/autotest_common.sh@1650 -- # local device=nvme1n3 00:06:23.863 15:54:21 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n3/queue/zoned ]] 00:06:23.863 15:54:21 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:06:23.863 15:54:21 -- spdk/autotest.sh@85 -- # (( 0 > 0 )) 00:06:23.863 15:54:21 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:06:23.863 15:54:21 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:06:23.863 15:54:21 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n1 00:06:23.863 15:54:21 -- scripts/common.sh@381 -- # local block=/dev/nvme0n1 pt 00:06:23.863 15:54:21 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:06:23.863 No valid GPT data, bailing 00:06:23.863 15:54:21 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:06:23.863 15:54:21 -- scripts/common.sh@394 -- # pt= 00:06:23.863 15:54:21 -- scripts/common.sh@395 -- # return 1 00:06:23.863 15:54:21 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:06:23.863 1+0 records in 00:06:23.863 1+0 records out 00:06:23.863 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00321274 s, 326 MB/s 00:06:23.863 15:54:21 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:06:23.863 15:54:21 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:06:23.863 15:54:21 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n1 00:06:23.863 15:54:21 -- scripts/common.sh@381 -- # local block=/dev/nvme1n1 pt 00:06:23.863 15:54:21 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n1 00:06:23.863 No valid GPT data, bailing 00:06:23.863 15:54:21 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n1 00:06:23.863 15:54:21 -- scripts/common.sh@394 -- # pt= 00:06:23.863 15:54:21 -- scripts/common.sh@395 -- # return 1 00:06:23.863 15:54:21 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n1 bs=1M count=1 00:06:23.863 1+0 records in 00:06:23.863 1+0 records out 00:06:23.863 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00498979 s, 210 MB/s 00:06:23.863 15:54:21 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:06:23.863 15:54:21 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:06:23.863 15:54:21 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n2 00:06:23.863 15:54:21 -- scripts/common.sh@381 -- # local block=/dev/nvme1n2 pt 00:06:23.863 15:54:21 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n2 00:06:23.863 No valid GPT data, bailing 00:06:23.863 15:54:21 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n2 00:06:23.863 15:54:21 -- scripts/common.sh@394 -- # pt= 00:06:23.863 15:54:21 -- scripts/common.sh@395 -- # return 1 00:06:23.863 15:54:21 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n2 bs=1M count=1 00:06:23.863 1+0 records in 00:06:23.863 1+0 records out 00:06:23.863 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00353308 s, 297 MB/s 00:06:23.863 15:54:21 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:06:23.863 15:54:21 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:06:23.864 15:54:21 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n3 00:06:23.864 15:54:21 -- scripts/common.sh@381 -- # local block=/dev/nvme1n3 pt 00:06:23.864 15:54:21 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n3 00:06:23.864 No valid GPT data, bailing 00:06:23.864 15:54:21 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n3 00:06:23.864 15:54:21 -- scripts/common.sh@394 -- # pt= 00:06:23.864 15:54:21 -- scripts/common.sh@395 -- # return 1 00:06:23.864 15:54:21 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n3 bs=1M count=1 00:06:23.864 1+0 records in 00:06:23.864 1+0 records out 00:06:23.864 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00467601 s, 224 MB/s 00:06:23.864 15:54:21 -- spdk/autotest.sh@105 -- # sync 00:06:23.864 15:54:21 -- spdk/autotest.sh@107 -- # xtrace_disable_per_cmd reap_spdk_processes 00:06:23.864 15:54:21 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:06:23.864 15:54:21 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:06:25.233 15:54:23 -- spdk/autotest.sh@111 -- # uname -s 00:06:25.233 15:54:23 -- spdk/autotest.sh@111 -- # [[ Linux == Linux ]] 00:06:25.233 15:54:23 -- spdk/autotest.sh@111 -- # [[ 0 -eq 1 ]] 00:06:25.233 15:54:23 -- spdk/autotest.sh@115 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:06:25.797 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:06:25.797 Hugepages 00:06:25.797 node hugesize free / total 00:06:25.797 node0 1048576kB 0 / 0 00:06:25.797 node0 2048kB 0 / 0 00:06:25.797 00:06:25.797 Type BDF Vendor Device NUMA Driver Device Block devices 00:06:25.797 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:06:25.797 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:06:26.054 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:06:26.054 15:54:24 -- spdk/autotest.sh@117 -- # uname -s 00:06:26.054 15:54:24 -- spdk/autotest.sh@117 -- # [[ Linux == Linux ]] 00:06:26.054 15:54:24 -- spdk/autotest.sh@119 -- # nvme_namespace_revert 00:06:26.054 15:54:24 -- common/autotest_common.sh@1516 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:06:26.618 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:06:26.618 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:06:26.618 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:06:26.618 15:54:24 -- common/autotest_common.sh@1517 -- # sleep 1 00:06:27.589 15:54:25 -- common/autotest_common.sh@1518 -- # bdfs=() 00:06:27.589 15:54:25 -- common/autotest_common.sh@1518 -- # local bdfs 00:06:27.589 15:54:25 -- common/autotest_common.sh@1520 -- # bdfs=($(get_nvme_bdfs)) 00:06:27.589 15:54:25 -- common/autotest_common.sh@1520 -- # get_nvme_bdfs 00:06:27.589 15:54:25 -- common/autotest_common.sh@1498 -- # bdfs=() 00:06:27.589 15:54:25 -- common/autotest_common.sh@1498 -- # local bdfs 00:06:27.589 15:54:25 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:06:27.589 15:54:25 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:06:27.589 15:54:25 -- common/autotest_common.sh@1499 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:06:27.589 15:54:25 -- common/autotest_common.sh@1500 -- # (( 2 == 0 )) 00:06:27.589 15:54:25 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:06:27.589 15:54:25 -- common/autotest_common.sh@1522 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:06:27.847 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:06:28.104 Waiting for block devices as requested 00:06:28.104 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:06:28.104 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:06:28.104 15:54:26 -- common/autotest_common.sh@1524 -- # for bdf in "${bdfs[@]}" 00:06:28.104 15:54:26 -- common/autotest_common.sh@1525 -- # get_nvme_ctrlr_from_bdf 0000:00:10.0 00:06:28.104 15:54:26 -- common/autotest_common.sh@1487 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:06:28.104 15:54:26 -- common/autotest_common.sh@1487 -- # grep 0000:00:10.0/nvme/nvme 00:06:28.104 15:54:26 -- common/autotest_common.sh@1487 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:06:28.104 15:54:26 -- common/autotest_common.sh@1488 -- # [[ -z /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 ]] 00:06:28.104 15:54:26 -- common/autotest_common.sh@1492 -- # basename /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:06:28.104 15:54:26 -- common/autotest_common.sh@1492 -- # printf '%s\n' nvme1 00:06:28.104 15:54:26 -- common/autotest_common.sh@1525 -- # nvme_ctrlr=/dev/nvme1 00:06:28.104 15:54:26 -- common/autotest_common.sh@1526 -- # [[ -z /dev/nvme1 ]] 00:06:28.104 15:54:26 -- common/autotest_common.sh@1531 -- # nvme id-ctrl /dev/nvme1 00:06:28.104 15:54:26 -- common/autotest_common.sh@1531 -- # grep oacs 00:06:28.104 15:54:26 -- common/autotest_common.sh@1531 -- # cut -d: -f2 00:06:28.104 15:54:26 -- common/autotest_common.sh@1531 -- # oacs=' 0x12a' 00:06:28.104 15:54:26 -- common/autotest_common.sh@1532 -- # oacs_ns_manage=8 00:06:28.104 15:54:26 -- common/autotest_common.sh@1534 -- # [[ 8 -ne 0 ]] 00:06:28.104 15:54:26 -- common/autotest_common.sh@1540 -- # nvme id-ctrl /dev/nvme1 00:06:28.104 15:54:26 -- common/autotest_common.sh@1540 -- # grep unvmcap 00:06:28.104 15:54:26 -- common/autotest_common.sh@1540 -- # cut -d: -f2 00:06:28.104 15:54:26 -- common/autotest_common.sh@1540 -- # unvmcap=' 0' 00:06:28.104 15:54:26 -- common/autotest_common.sh@1541 -- # [[ 0 -eq 0 ]] 00:06:28.104 15:54:26 -- common/autotest_common.sh@1543 -- # continue 00:06:28.104 15:54:26 -- common/autotest_common.sh@1524 -- # for bdf in "${bdfs[@]}" 00:06:28.104 15:54:26 -- common/autotest_common.sh@1525 -- # get_nvme_ctrlr_from_bdf 0000:00:11.0 00:06:28.104 15:54:26 -- common/autotest_common.sh@1487 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:06:28.104 15:54:26 -- common/autotest_common.sh@1487 -- # grep 0000:00:11.0/nvme/nvme 00:06:28.104 15:54:26 -- common/autotest_common.sh@1487 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:06:28.104 15:54:26 -- common/autotest_common.sh@1488 -- # [[ -z /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 ]] 00:06:28.104 15:54:26 -- common/autotest_common.sh@1492 -- # basename /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:06:28.104 15:54:26 -- common/autotest_common.sh@1492 -- # printf '%s\n' nvme0 00:06:28.104 15:54:26 -- common/autotest_common.sh@1525 -- # nvme_ctrlr=/dev/nvme0 00:06:28.104 15:54:26 -- common/autotest_common.sh@1526 -- # [[ -z /dev/nvme0 ]] 00:06:28.104 15:54:26 -- common/autotest_common.sh@1531 -- # grep oacs 00:06:28.104 15:54:26 -- common/autotest_common.sh@1531 -- # nvme id-ctrl /dev/nvme0 00:06:28.104 15:54:26 -- common/autotest_common.sh@1531 -- # cut -d: -f2 00:06:28.104 15:54:26 -- common/autotest_common.sh@1531 -- # oacs=' 0x12a' 00:06:28.104 15:54:26 -- common/autotest_common.sh@1532 -- # oacs_ns_manage=8 00:06:28.104 15:54:26 -- common/autotest_common.sh@1534 -- # [[ 8 -ne 0 ]] 00:06:28.105 15:54:26 -- common/autotest_common.sh@1540 -- # grep unvmcap 00:06:28.105 15:54:26 -- common/autotest_common.sh@1540 -- # nvme id-ctrl /dev/nvme0 00:06:28.105 15:54:26 -- common/autotest_common.sh@1540 -- # cut -d: -f2 00:06:28.105 15:54:26 -- common/autotest_common.sh@1540 -- # unvmcap=' 0' 00:06:28.105 15:54:26 -- common/autotest_common.sh@1541 -- # [[ 0 -eq 0 ]] 00:06:28.105 15:54:26 -- common/autotest_common.sh@1543 -- # continue 00:06:28.105 15:54:26 -- spdk/autotest.sh@122 -- # timing_exit pre_cleanup 00:06:28.105 15:54:26 -- common/autotest_common.sh@732 -- # xtrace_disable 00:06:28.105 15:54:26 -- common/autotest_common.sh@10 -- # set +x 00:06:28.364 15:54:26 -- spdk/autotest.sh@125 -- # timing_enter afterboot 00:06:28.364 15:54:26 -- common/autotest_common.sh@726 -- # xtrace_disable 00:06:28.364 15:54:26 -- common/autotest_common.sh@10 -- # set +x 00:06:28.364 15:54:26 -- spdk/autotest.sh@126 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:06:28.621 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:06:28.877 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:06:28.877 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:06:28.877 15:54:27 -- spdk/autotest.sh@127 -- # timing_exit afterboot 00:06:28.877 15:54:27 -- common/autotest_common.sh@732 -- # xtrace_disable 00:06:28.877 15:54:27 -- common/autotest_common.sh@10 -- # set +x 00:06:28.877 15:54:27 -- spdk/autotest.sh@131 -- # opal_revert_cleanup 00:06:28.877 15:54:27 -- common/autotest_common.sh@1578 -- # mapfile -t bdfs 00:06:28.877 15:54:27 -- common/autotest_common.sh@1578 -- # get_nvme_bdfs_by_id 0x0a54 00:06:28.877 15:54:27 -- common/autotest_common.sh@1563 -- # bdfs=() 00:06:28.877 15:54:27 -- common/autotest_common.sh@1563 -- # _bdfs=() 00:06:28.877 15:54:27 -- common/autotest_common.sh@1563 -- # local bdfs _bdfs 00:06:28.877 15:54:27 -- common/autotest_common.sh@1564 -- # _bdfs=($(get_nvme_bdfs)) 00:06:28.877 15:54:27 -- common/autotest_common.sh@1564 -- # get_nvme_bdfs 00:06:28.877 15:54:27 -- common/autotest_common.sh@1498 -- # bdfs=() 00:06:28.877 15:54:27 -- common/autotest_common.sh@1498 -- # local bdfs 00:06:28.877 15:54:27 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:06:28.877 15:54:27 -- common/autotest_common.sh@1499 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:06:28.877 15:54:27 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:06:29.135 15:54:27 -- common/autotest_common.sh@1500 -- # (( 2 == 0 )) 00:06:29.135 15:54:27 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:06:29.135 15:54:27 -- common/autotest_common.sh@1565 -- # for bdf in "${_bdfs[@]}" 00:06:29.135 15:54:27 -- common/autotest_common.sh@1566 -- # cat /sys/bus/pci/devices/0000:00:10.0/device 00:06:29.135 15:54:27 -- common/autotest_common.sh@1566 -- # device=0x0010 00:06:29.135 15:54:27 -- common/autotest_common.sh@1567 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:06:29.135 15:54:27 -- common/autotest_common.sh@1565 -- # for bdf in "${_bdfs[@]}" 00:06:29.135 15:54:27 -- common/autotest_common.sh@1566 -- # cat /sys/bus/pci/devices/0000:00:11.0/device 00:06:29.135 15:54:27 -- common/autotest_common.sh@1566 -- # device=0x0010 00:06:29.135 15:54:27 -- common/autotest_common.sh@1567 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:06:29.135 15:54:27 -- common/autotest_common.sh@1572 -- # (( 0 > 0 )) 00:06:29.135 15:54:27 -- common/autotest_common.sh@1572 -- # return 0 00:06:29.135 15:54:27 -- common/autotest_common.sh@1579 -- # [[ -z '' ]] 00:06:29.135 15:54:27 -- common/autotest_common.sh@1580 -- # return 0 00:06:29.135 15:54:27 -- spdk/autotest.sh@137 -- # '[' 0 -eq 1 ']' 00:06:29.135 15:54:27 -- spdk/autotest.sh@141 -- # '[' 1 -eq 1 ']' 00:06:29.135 15:54:27 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:06:29.135 15:54:27 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:06:29.135 15:54:27 -- spdk/autotest.sh@149 -- # timing_enter lib 00:06:29.135 15:54:27 -- common/autotest_common.sh@726 -- # xtrace_disable 00:06:29.135 15:54:27 -- common/autotest_common.sh@10 -- # set +x 00:06:29.135 15:54:27 -- spdk/autotest.sh@151 -- # [[ 0 -eq 1 ]] 00:06:29.135 15:54:27 -- spdk/autotest.sh@155 -- # run_test env /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:06:29.135 15:54:27 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:29.135 15:54:27 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:29.135 15:54:27 -- common/autotest_common.sh@10 -- # set +x 00:06:29.135 ************************************ 00:06:29.135 START TEST env 00:06:29.135 ************************************ 00:06:29.135 15:54:27 env -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:06:29.135 * Looking for test storage... 00:06:29.135 * Found test storage at /home/vagrant/spdk_repo/spdk/test/env 00:06:29.135 15:54:27 env -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:29.135 15:54:27 env -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:29.135 15:54:27 env -- common/autotest_common.sh@1693 -- # lcov --version 00:06:29.135 15:54:27 env -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:29.135 15:54:27 env -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:29.135 15:54:27 env -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:29.135 15:54:27 env -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:29.135 15:54:27 env -- scripts/common.sh@336 -- # IFS=.-: 00:06:29.135 15:54:27 env -- scripts/common.sh@336 -- # read -ra ver1 00:06:29.135 15:54:27 env -- scripts/common.sh@337 -- # IFS=.-: 00:06:29.135 15:54:27 env -- scripts/common.sh@337 -- # read -ra ver2 00:06:29.135 15:54:27 env -- scripts/common.sh@338 -- # local 'op=<' 00:06:29.135 15:54:27 env -- scripts/common.sh@340 -- # ver1_l=2 00:06:29.135 15:54:27 env -- scripts/common.sh@341 -- # ver2_l=1 00:06:29.135 15:54:27 env -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:29.135 15:54:27 env -- scripts/common.sh@344 -- # case "$op" in 00:06:29.135 15:54:27 env -- scripts/common.sh@345 -- # : 1 00:06:29.135 15:54:27 env -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:29.135 15:54:27 env -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:29.135 15:54:27 env -- scripts/common.sh@365 -- # decimal 1 00:06:29.135 15:54:27 env -- scripts/common.sh@353 -- # local d=1 00:06:29.135 15:54:27 env -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:29.135 15:54:27 env -- scripts/common.sh@355 -- # echo 1 00:06:29.135 15:54:27 env -- scripts/common.sh@365 -- # ver1[v]=1 00:06:29.135 15:54:27 env -- scripts/common.sh@366 -- # decimal 2 00:06:29.135 15:54:27 env -- scripts/common.sh@353 -- # local d=2 00:06:29.135 15:54:27 env -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:29.135 15:54:27 env -- scripts/common.sh@355 -- # echo 2 00:06:29.135 15:54:27 env -- scripts/common.sh@366 -- # ver2[v]=2 00:06:29.135 15:54:27 env -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:29.135 15:54:27 env -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:29.135 15:54:27 env -- scripts/common.sh@368 -- # return 0 00:06:29.135 15:54:27 env -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:29.135 15:54:27 env -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:29.135 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:29.135 --rc genhtml_branch_coverage=1 00:06:29.135 --rc genhtml_function_coverage=1 00:06:29.135 --rc genhtml_legend=1 00:06:29.135 --rc geninfo_all_blocks=1 00:06:29.135 --rc geninfo_unexecuted_blocks=1 00:06:29.135 00:06:29.135 ' 00:06:29.135 15:54:27 env -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:29.135 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:29.135 --rc genhtml_branch_coverage=1 00:06:29.135 --rc genhtml_function_coverage=1 00:06:29.135 --rc genhtml_legend=1 00:06:29.135 --rc geninfo_all_blocks=1 00:06:29.135 --rc geninfo_unexecuted_blocks=1 00:06:29.135 00:06:29.135 ' 00:06:29.135 15:54:27 env -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:29.135 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:29.135 --rc genhtml_branch_coverage=1 00:06:29.135 --rc genhtml_function_coverage=1 00:06:29.135 --rc genhtml_legend=1 00:06:29.135 --rc geninfo_all_blocks=1 00:06:29.135 --rc geninfo_unexecuted_blocks=1 00:06:29.135 00:06:29.135 ' 00:06:29.135 15:54:27 env -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:29.135 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:29.135 --rc genhtml_branch_coverage=1 00:06:29.135 --rc genhtml_function_coverage=1 00:06:29.135 --rc genhtml_legend=1 00:06:29.135 --rc geninfo_all_blocks=1 00:06:29.135 --rc geninfo_unexecuted_blocks=1 00:06:29.135 00:06:29.135 ' 00:06:29.135 15:54:27 env -- env/env.sh@10 -- # run_test env_memory /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:06:29.135 15:54:27 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:29.135 15:54:27 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:29.135 15:54:27 env -- common/autotest_common.sh@10 -- # set +x 00:06:29.135 ************************************ 00:06:29.135 START TEST env_memory 00:06:29.135 ************************************ 00:06:29.135 15:54:27 env.env_memory -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:06:29.135 00:06:29.135 00:06:29.135 CUnit - A unit testing framework for C - Version 2.1-3 00:06:29.135 http://cunit.sourceforge.net/ 00:06:29.135 00:06:29.135 00:06:29.135 Suite: memory 00:06:29.135 Test: alloc and free memory map ...[2024-11-20 15:54:27.370982] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:06:29.393 passed 00:06:29.393 Test: mem map translation ...[2024-11-20 15:54:27.410581] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:06:29.393 [2024-11-20 15:54:27.410641] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:06:29.393 [2024-11-20 15:54:27.410710] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 589:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:06:29.393 [2024-11-20 15:54:27.410727] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 605:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:06:29.393 passed 00:06:29.393 Test: mem map registration ...[2024-11-20 15:54:27.480596] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=200000 len=1234 00:06:29.393 [2024-11-20 15:54:27.480698] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=4d2 len=2097152 00:06:29.393 passed 00:06:29.393 Test: mem map adjacent registrations ...passed 00:06:29.393 00:06:29.393 Run Summary: Type Total Ran Passed Failed Inactive 00:06:29.393 suites 1 1 n/a 0 0 00:06:29.393 tests 4 4 4 0 0 00:06:29.393 asserts 152 152 152 0 n/a 00:06:29.393 00:06:29.393 Elapsed time = 0.236 seconds 00:06:29.393 00:06:29.393 real 0m0.265s 00:06:29.393 user 0m0.238s 00:06:29.393 sys 0m0.021s 00:06:29.393 15:54:27 env.env_memory -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:29.393 15:54:27 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:06:29.393 ************************************ 00:06:29.393 END TEST env_memory 00:06:29.393 ************************************ 00:06:29.393 15:54:27 env -- env/env.sh@11 -- # run_test env_vtophys /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:06:29.393 15:54:27 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:29.393 15:54:27 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:29.393 15:54:27 env -- common/autotest_common.sh@10 -- # set +x 00:06:29.393 ************************************ 00:06:29.393 START TEST env_vtophys 00:06:29.393 ************************************ 00:06:29.393 15:54:27 env.env_vtophys -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:06:29.649 EAL: lib.eal log level changed from notice to debug 00:06:29.649 EAL: Detected lcore 0 as core 0 on socket 0 00:06:29.649 EAL: Detected lcore 1 as core 0 on socket 0 00:06:29.649 EAL: Detected lcore 2 as core 0 on socket 0 00:06:29.649 EAL: Detected lcore 3 as core 0 on socket 0 00:06:29.649 EAL: Detected lcore 4 as core 0 on socket 0 00:06:29.649 EAL: Detected lcore 5 as core 0 on socket 0 00:06:29.649 EAL: Detected lcore 6 as core 0 on socket 0 00:06:29.649 EAL: Detected lcore 7 as core 0 on socket 0 00:06:29.649 EAL: Detected lcore 8 as core 0 on socket 0 00:06:29.649 EAL: Detected lcore 9 as core 0 on socket 0 00:06:29.649 EAL: Maximum logical cores by configuration: 128 00:06:29.649 EAL: Detected CPU lcores: 10 00:06:29.649 EAL: Detected NUMA nodes: 1 00:06:29.649 EAL: Checking presence of .so 'librte_eal.so.24.1' 00:06:29.649 EAL: Detected shared linkage of DPDK 00:06:29.649 EAL: No shared files mode enabled, IPC will be disabled 00:06:29.649 EAL: Selected IOVA mode 'PA' 00:06:29.649 EAL: Probing VFIO support... 00:06:29.649 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:06:29.649 EAL: VFIO modules not loaded, skipping VFIO support... 00:06:29.649 EAL: Ask a virtual area of 0x2e000 bytes 00:06:29.649 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:06:29.649 EAL: Setting up physically contiguous memory... 00:06:29.649 EAL: Setting maximum number of open files to 524288 00:06:29.649 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:06:29.649 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:06:29.649 EAL: Ask a virtual area of 0x61000 bytes 00:06:29.649 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:06:29.649 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:06:29.649 EAL: Ask a virtual area of 0x400000000 bytes 00:06:29.649 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:06:29.649 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:06:29.649 EAL: Ask a virtual area of 0x61000 bytes 00:06:29.649 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:06:29.649 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:06:29.649 EAL: Ask a virtual area of 0x400000000 bytes 00:06:29.649 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:06:29.649 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:06:29.649 EAL: Ask a virtual area of 0x61000 bytes 00:06:29.649 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:06:29.649 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:06:29.649 EAL: Ask a virtual area of 0x400000000 bytes 00:06:29.649 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:06:29.649 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:06:29.649 EAL: Ask a virtual area of 0x61000 bytes 00:06:29.649 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:06:29.649 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:06:29.649 EAL: Ask a virtual area of 0x400000000 bytes 00:06:29.649 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:06:29.649 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:06:29.649 EAL: Hugepages will be freed exactly as allocated. 00:06:29.649 EAL: No shared files mode enabled, IPC is disabled 00:06:29.649 EAL: No shared files mode enabled, IPC is disabled 00:06:29.649 EAL: TSC frequency is ~2600000 KHz 00:06:29.649 EAL: Main lcore 0 is ready (tid=7fcb7f8b5a40;cpuset=[0]) 00:06:29.649 EAL: Trying to obtain current memory policy. 00:06:29.649 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:29.649 EAL: Restoring previous memory policy: 0 00:06:29.649 EAL: request: mp_malloc_sync 00:06:29.649 EAL: No shared files mode enabled, IPC is disabled 00:06:29.649 EAL: Heap on socket 0 was expanded by 2MB 00:06:29.649 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:06:29.649 EAL: No PCI address specified using 'addr=' in: bus=pci 00:06:29.649 EAL: Mem event callback 'spdk:(nil)' registered 00:06:29.649 EAL: Module /sys/module/vfio_pci not found! error 2 (No such file or directory) 00:06:29.649 00:06:29.649 00:06:29.649 CUnit - A unit testing framework for C - Version 2.1-3 00:06:29.649 http://cunit.sourceforge.net/ 00:06:29.649 00:06:29.649 00:06:29.649 Suite: components_suite 00:06:29.905 Test: vtophys_malloc_test ...passed 00:06:29.905 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:06:29.905 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:29.905 EAL: Restoring previous memory policy: 4 00:06:29.905 EAL: Calling mem event callback 'spdk:(nil)' 00:06:29.905 EAL: request: mp_malloc_sync 00:06:29.905 EAL: No shared files mode enabled, IPC is disabled 00:06:29.905 EAL: Heap on socket 0 was expanded by 4MB 00:06:29.905 EAL: Calling mem event callback 'spdk:(nil)' 00:06:29.905 EAL: request: mp_malloc_sync 00:06:29.905 EAL: No shared files mode enabled, IPC is disabled 00:06:29.905 EAL: Heap on socket 0 was shrunk by 4MB 00:06:29.905 EAL: Trying to obtain current memory policy. 00:06:29.905 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:29.905 EAL: Restoring previous memory policy: 4 00:06:29.905 EAL: Calling mem event callback 'spdk:(nil)' 00:06:29.905 EAL: request: mp_malloc_sync 00:06:29.905 EAL: No shared files mode enabled, IPC is disabled 00:06:29.905 EAL: Heap on socket 0 was expanded by 6MB 00:06:29.905 EAL: Calling mem event callback 'spdk:(nil)' 00:06:29.905 EAL: request: mp_malloc_sync 00:06:29.905 EAL: No shared files mode enabled, IPC is disabled 00:06:29.905 EAL: Heap on socket 0 was shrunk by 6MB 00:06:29.905 EAL: Trying to obtain current memory policy. 00:06:29.905 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:29.905 EAL: Restoring previous memory policy: 4 00:06:29.905 EAL: Calling mem event callback 'spdk:(nil)' 00:06:29.905 EAL: request: mp_malloc_sync 00:06:29.905 EAL: No shared files mode enabled, IPC is disabled 00:06:29.905 EAL: Heap on socket 0 was expanded by 10MB 00:06:29.905 EAL: Calling mem event callback 'spdk:(nil)' 00:06:29.905 EAL: request: mp_malloc_sync 00:06:29.905 EAL: No shared files mode enabled, IPC is disabled 00:06:29.905 EAL: Heap on socket 0 was shrunk by 10MB 00:06:30.162 EAL: Trying to obtain current memory policy. 00:06:30.162 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:30.162 EAL: Restoring previous memory policy: 4 00:06:30.162 EAL: Calling mem event callback 'spdk:(nil)' 00:06:30.162 EAL: request: mp_malloc_sync 00:06:30.162 EAL: No shared files mode enabled, IPC is disabled 00:06:30.162 EAL: Heap on socket 0 was expanded by 18MB 00:06:30.162 EAL: Calling mem event callback 'spdk:(nil)' 00:06:30.162 EAL: request: mp_malloc_sync 00:06:30.162 EAL: No shared files mode enabled, IPC is disabled 00:06:30.162 EAL: Heap on socket 0 was shrunk by 18MB 00:06:30.162 EAL: Trying to obtain current memory policy. 00:06:30.162 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:30.162 EAL: Restoring previous memory policy: 4 00:06:30.162 EAL: Calling mem event callback 'spdk:(nil)' 00:06:30.162 EAL: request: mp_malloc_sync 00:06:30.162 EAL: No shared files mode enabled, IPC is disabled 00:06:30.162 EAL: Heap on socket 0 was expanded by 34MB 00:06:30.162 EAL: Calling mem event callback 'spdk:(nil)' 00:06:30.162 EAL: request: mp_malloc_sync 00:06:30.162 EAL: No shared files mode enabled, IPC is disabled 00:06:30.162 EAL: Heap on socket 0 was shrunk by 34MB 00:06:30.162 EAL: Trying to obtain current memory policy. 00:06:30.162 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:30.162 EAL: Restoring previous memory policy: 4 00:06:30.162 EAL: Calling mem event callback 'spdk:(nil)' 00:06:30.162 EAL: request: mp_malloc_sync 00:06:30.162 EAL: No shared files mode enabled, IPC is disabled 00:06:30.162 EAL: Heap on socket 0 was expanded by 66MB 00:06:30.162 EAL: Calling mem event callback 'spdk:(nil)' 00:06:30.162 EAL: request: mp_malloc_sync 00:06:30.162 EAL: No shared files mode enabled, IPC is disabled 00:06:30.162 EAL: Heap on socket 0 was shrunk by 66MB 00:06:30.419 EAL: Trying to obtain current memory policy. 00:06:30.419 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:30.419 EAL: Restoring previous memory policy: 4 00:06:30.419 EAL: Calling mem event callback 'spdk:(nil)' 00:06:30.419 EAL: request: mp_malloc_sync 00:06:30.419 EAL: No shared files mode enabled, IPC is disabled 00:06:30.419 EAL: Heap on socket 0 was expanded by 130MB 00:06:30.419 EAL: Calling mem event callback 'spdk:(nil)' 00:06:30.419 EAL: request: mp_malloc_sync 00:06:30.419 EAL: No shared files mode enabled, IPC is disabled 00:06:30.419 EAL: Heap on socket 0 was shrunk by 130MB 00:06:30.676 EAL: Trying to obtain current memory policy. 00:06:30.676 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:30.676 EAL: Restoring previous memory policy: 4 00:06:30.676 EAL: Calling mem event callback 'spdk:(nil)' 00:06:30.676 EAL: request: mp_malloc_sync 00:06:30.676 EAL: No shared files mode enabled, IPC is disabled 00:06:30.676 EAL: Heap on socket 0 was expanded by 258MB 00:06:30.933 EAL: Calling mem event callback 'spdk:(nil)' 00:06:30.933 EAL: request: mp_malloc_sync 00:06:30.933 EAL: No shared files mode enabled, IPC is disabled 00:06:30.933 EAL: Heap on socket 0 was shrunk by 258MB 00:06:31.190 EAL: Trying to obtain current memory policy. 00:06:31.190 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:31.190 EAL: Restoring previous memory policy: 4 00:06:31.190 EAL: Calling mem event callback 'spdk:(nil)' 00:06:31.190 EAL: request: mp_malloc_sync 00:06:31.190 EAL: No shared files mode enabled, IPC is disabled 00:06:31.190 EAL: Heap on socket 0 was expanded by 514MB 00:06:31.862 EAL: Calling mem event callback 'spdk:(nil)' 00:06:31.862 EAL: request: mp_malloc_sync 00:06:31.862 EAL: No shared files mode enabled, IPC is disabled 00:06:31.862 EAL: Heap on socket 0 was shrunk by 514MB 00:06:32.427 EAL: Trying to obtain current memory policy. 00:06:32.427 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:32.685 EAL: Restoring previous memory policy: 4 00:06:32.685 EAL: Calling mem event callback 'spdk:(nil)' 00:06:32.685 EAL: request: mp_malloc_sync 00:06:32.685 EAL: No shared files mode enabled, IPC is disabled 00:06:32.685 EAL: Heap on socket 0 was expanded by 1026MB 00:06:33.626 EAL: Calling mem event callback 'spdk:(nil)' 00:06:33.889 EAL: request: mp_malloc_sync 00:06:33.889 EAL: No shared files mode enabled, IPC is disabled 00:06:33.889 EAL: Heap on socket 0 was shrunk by 1026MB 00:06:34.919 passed 00:06:34.919 00:06:34.919 Run Summary: Type Total Ran Passed Failed Inactive 00:06:34.919 suites 1 1 n/a 0 0 00:06:34.919 tests 2 2 2 0 0 00:06:34.919 asserts 5691 5691 5691 0 n/a 00:06:34.919 00:06:34.919 Elapsed time = 5.133 seconds 00:06:34.919 EAL: Calling mem event callback 'spdk:(nil)' 00:06:34.919 EAL: request: mp_malloc_sync 00:06:34.919 EAL: No shared files mode enabled, IPC is disabled 00:06:34.919 EAL: Heap on socket 0 was shrunk by 2MB 00:06:34.919 EAL: No shared files mode enabled, IPC is disabled 00:06:34.919 EAL: No shared files mode enabled, IPC is disabled 00:06:34.919 EAL: No shared files mode enabled, IPC is disabled 00:06:34.919 00:06:34.919 real 0m5.389s 00:06:34.919 user 0m4.613s 00:06:34.919 sys 0m0.632s 00:06:34.919 15:54:33 env.env_vtophys -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:34.919 15:54:33 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:06:34.919 ************************************ 00:06:34.919 END TEST env_vtophys 00:06:34.919 ************************************ 00:06:34.919 15:54:33 env -- env/env.sh@12 -- # run_test env_pci /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:06:34.919 15:54:33 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:34.919 15:54:33 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:34.919 15:54:33 env -- common/autotest_common.sh@10 -- # set +x 00:06:34.919 ************************************ 00:06:34.919 START TEST env_pci 00:06:34.919 ************************************ 00:06:34.919 15:54:33 env.env_pci -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:06:34.919 00:06:34.919 00:06:34.919 CUnit - A unit testing framework for C - Version 2.1-3 00:06:34.919 http://cunit.sourceforge.net/ 00:06:34.919 00:06:34.919 00:06:34.919 Suite: pci 00:06:34.919 Test: pci_hook ...[2024-11-20 15:54:33.078552] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/pci.c:1117:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 55954 has claimed it 00:06:34.919 passed 00:06:34.919 00:06:34.919 EAL: Cannot find device (10000:00:01.0) 00:06:34.919 EAL: Failed to attach device on primary process 00:06:34.919 Run Summary: Type Total Ran Passed Failed Inactive 00:06:34.919 suites 1 1 n/a 0 0 00:06:34.919 tests 1 1 1 0 0 00:06:34.919 asserts 25 25 25 0 n/a 00:06:34.919 00:06:34.919 Elapsed time = 0.004 seconds 00:06:34.919 00:06:34.919 real 0m0.060s 00:06:34.919 user 0m0.026s 00:06:34.919 sys 0m0.033s 00:06:34.919 15:54:33 env.env_pci -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:34.919 15:54:33 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:06:34.919 ************************************ 00:06:34.919 END TEST env_pci 00:06:34.919 ************************************ 00:06:35.191 15:54:33 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:06:35.191 15:54:33 env -- env/env.sh@15 -- # uname 00:06:35.191 15:54:33 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:06:35.191 15:54:33 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:06:35.191 15:54:33 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:06:35.191 15:54:33 env -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:06:35.191 15:54:33 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:35.191 15:54:33 env -- common/autotest_common.sh@10 -- # set +x 00:06:35.191 ************************************ 00:06:35.191 START TEST env_dpdk_post_init 00:06:35.191 ************************************ 00:06:35.191 15:54:33 env.env_dpdk_post_init -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:06:35.191 EAL: Detected CPU lcores: 10 00:06:35.191 EAL: Detected NUMA nodes: 1 00:06:35.191 EAL: Detected shared linkage of DPDK 00:06:35.191 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:06:35.191 EAL: Selected IOVA mode 'PA' 00:06:35.191 TELEMETRY: No legacy callbacks, legacy socket not created 00:06:35.191 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:10.0 (socket -1) 00:06:35.191 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:11.0 (socket -1) 00:06:35.191 Starting DPDK initialization... 00:06:35.191 Starting SPDK post initialization... 00:06:35.191 SPDK NVMe probe 00:06:35.191 Attaching to 0000:00:10.0 00:06:35.191 Attaching to 0000:00:11.0 00:06:35.191 Attached to 0000:00:10.0 00:06:35.191 Attached to 0000:00:11.0 00:06:35.191 Cleaning up... 00:06:35.191 00:06:35.191 real 0m0.236s 00:06:35.191 user 0m0.077s 00:06:35.191 sys 0m0.059s 00:06:35.191 15:54:33 env.env_dpdk_post_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:35.191 15:54:33 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:06:35.191 ************************************ 00:06:35.191 END TEST env_dpdk_post_init 00:06:35.191 ************************************ 00:06:35.449 15:54:33 env -- env/env.sh@26 -- # uname 00:06:35.449 15:54:33 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:06:35.449 15:54:33 env -- env/env.sh@29 -- # run_test env_mem_callbacks /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:06:35.449 15:54:33 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:35.449 15:54:33 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:35.449 15:54:33 env -- common/autotest_common.sh@10 -- # set +x 00:06:35.449 ************************************ 00:06:35.449 START TEST env_mem_callbacks 00:06:35.449 ************************************ 00:06:35.449 15:54:33 env.env_mem_callbacks -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:06:35.449 EAL: Detected CPU lcores: 10 00:06:35.449 EAL: Detected NUMA nodes: 1 00:06:35.449 EAL: Detected shared linkage of DPDK 00:06:35.449 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:06:35.449 EAL: Selected IOVA mode 'PA' 00:06:35.449 TELEMETRY: No legacy callbacks, legacy socket not created 00:06:35.449 00:06:35.449 00:06:35.449 CUnit - A unit testing framework for C - Version 2.1-3 00:06:35.449 http://cunit.sourceforge.net/ 00:06:35.449 00:06:35.449 00:06:35.449 Suite: memory 00:06:35.449 Test: test ... 00:06:35.449 register 0x200000200000 2097152 00:06:35.449 malloc 3145728 00:06:35.449 register 0x200000400000 4194304 00:06:35.449 buf 0x2000004fffc0 len 3145728 PASSED 00:06:35.449 malloc 64 00:06:35.449 buf 0x2000004ffec0 len 64 PASSED 00:06:35.449 malloc 4194304 00:06:35.449 register 0x200000800000 6291456 00:06:35.449 buf 0x2000009fffc0 len 4194304 PASSED 00:06:35.449 free 0x2000004fffc0 3145728 00:06:35.449 free 0x2000004ffec0 64 00:06:35.449 unregister 0x200000400000 4194304 PASSED 00:06:35.449 free 0x2000009fffc0 4194304 00:06:35.449 unregister 0x200000800000 6291456 PASSED 00:06:35.449 malloc 8388608 00:06:35.449 register 0x200000400000 10485760 00:06:35.449 buf 0x2000005fffc0 len 8388608 PASSED 00:06:35.449 free 0x2000005fffc0 8388608 00:06:35.449 unregister 0x200000400000 10485760 PASSED 00:06:35.449 passed 00:06:35.449 00:06:35.449 Run Summary: Type Total Ran Passed Failed Inactive 00:06:35.449 suites 1 1 n/a 0 0 00:06:35.449 tests 1 1 1 0 0 00:06:35.449 asserts 15 15 15 0 n/a 00:06:35.449 00:06:35.449 Elapsed time = 0.040 seconds 00:06:35.449 00:06:35.449 real 0m0.205s 00:06:35.449 user 0m0.062s 00:06:35.449 sys 0m0.042s 00:06:35.449 15:54:33 env.env_mem_callbacks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:35.449 15:54:33 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:06:35.449 ************************************ 00:06:35.449 END TEST env_mem_callbacks 00:06:35.449 ************************************ 00:06:35.711 00:06:35.711 real 0m6.532s 00:06:35.711 user 0m5.163s 00:06:35.711 sys 0m0.990s 00:06:35.711 15:54:33 env -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:35.711 15:54:33 env -- common/autotest_common.sh@10 -- # set +x 00:06:35.711 ************************************ 00:06:35.711 END TEST env 00:06:35.711 ************************************ 00:06:35.711 15:54:33 -- spdk/autotest.sh@156 -- # run_test rpc /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:06:35.711 15:54:33 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:35.711 15:54:33 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:35.711 15:54:33 -- common/autotest_common.sh@10 -- # set +x 00:06:35.711 ************************************ 00:06:35.711 START TEST rpc 00:06:35.711 ************************************ 00:06:35.711 15:54:33 rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:06:35.711 * Looking for test storage... 00:06:35.711 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:06:35.711 15:54:33 rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:35.711 15:54:33 rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:06:35.711 15:54:33 rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:35.711 15:54:33 rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:35.711 15:54:33 rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:35.711 15:54:33 rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:35.711 15:54:33 rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:35.711 15:54:33 rpc -- scripts/common.sh@336 -- # IFS=.-: 00:06:35.711 15:54:33 rpc -- scripts/common.sh@336 -- # read -ra ver1 00:06:35.711 15:54:33 rpc -- scripts/common.sh@337 -- # IFS=.-: 00:06:35.711 15:54:33 rpc -- scripts/common.sh@337 -- # read -ra ver2 00:06:35.711 15:54:33 rpc -- scripts/common.sh@338 -- # local 'op=<' 00:06:35.711 15:54:33 rpc -- scripts/common.sh@340 -- # ver1_l=2 00:06:35.711 15:54:33 rpc -- scripts/common.sh@341 -- # ver2_l=1 00:06:35.711 15:54:33 rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:35.711 15:54:33 rpc -- scripts/common.sh@344 -- # case "$op" in 00:06:35.711 15:54:33 rpc -- scripts/common.sh@345 -- # : 1 00:06:35.711 15:54:33 rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:35.711 15:54:33 rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:35.711 15:54:33 rpc -- scripts/common.sh@365 -- # decimal 1 00:06:35.711 15:54:33 rpc -- scripts/common.sh@353 -- # local d=1 00:06:35.711 15:54:33 rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:35.711 15:54:33 rpc -- scripts/common.sh@355 -- # echo 1 00:06:35.711 15:54:33 rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:06:35.711 15:54:33 rpc -- scripts/common.sh@366 -- # decimal 2 00:06:35.711 15:54:33 rpc -- scripts/common.sh@353 -- # local d=2 00:06:35.711 15:54:33 rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:35.711 15:54:33 rpc -- scripts/common.sh@355 -- # echo 2 00:06:35.711 15:54:33 rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:06:35.711 15:54:33 rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:35.711 15:54:33 rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:35.711 15:54:33 rpc -- scripts/common.sh@368 -- # return 0 00:06:35.711 15:54:33 rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:35.711 15:54:33 rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:35.711 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:35.711 --rc genhtml_branch_coverage=1 00:06:35.711 --rc genhtml_function_coverage=1 00:06:35.711 --rc genhtml_legend=1 00:06:35.711 --rc geninfo_all_blocks=1 00:06:35.711 --rc geninfo_unexecuted_blocks=1 00:06:35.711 00:06:35.711 ' 00:06:35.711 15:54:33 rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:35.711 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:35.711 --rc genhtml_branch_coverage=1 00:06:35.711 --rc genhtml_function_coverage=1 00:06:35.711 --rc genhtml_legend=1 00:06:35.711 --rc geninfo_all_blocks=1 00:06:35.711 --rc geninfo_unexecuted_blocks=1 00:06:35.711 00:06:35.711 ' 00:06:35.711 15:54:33 rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:35.711 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:35.711 --rc genhtml_branch_coverage=1 00:06:35.711 --rc genhtml_function_coverage=1 00:06:35.711 --rc genhtml_legend=1 00:06:35.711 --rc geninfo_all_blocks=1 00:06:35.711 --rc geninfo_unexecuted_blocks=1 00:06:35.711 00:06:35.711 ' 00:06:35.711 15:54:33 rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:35.711 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:35.711 --rc genhtml_branch_coverage=1 00:06:35.711 --rc genhtml_function_coverage=1 00:06:35.711 --rc genhtml_legend=1 00:06:35.711 --rc geninfo_all_blocks=1 00:06:35.711 --rc geninfo_unexecuted_blocks=1 00:06:35.711 00:06:35.711 ' 00:06:35.711 15:54:33 rpc -- rpc/rpc.sh@65 -- # spdk_pid=56076 00:06:35.711 15:54:33 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:35.711 15:54:33 rpc -- rpc/rpc.sh@67 -- # waitforlisten 56076 00:06:35.711 15:54:33 rpc -- common/autotest_common.sh@835 -- # '[' -z 56076 ']' 00:06:35.711 15:54:33 rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:35.711 15:54:33 rpc -- rpc/rpc.sh@64 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -e bdev 00:06:35.711 15:54:33 rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:35.711 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:35.711 15:54:33 rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:35.711 15:54:33 rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:35.711 15:54:33 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:35.971 [2024-11-20 15:54:33.980883] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:06:35.971 [2024-11-20 15:54:33.981004] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid56076 ] 00:06:35.971 [2024-11-20 15:54:34.134212] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:36.233 [2024-11-20 15:54:34.233655] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:06:36.233 [2024-11-20 15:54:34.233717] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 56076' to capture a snapshot of events at runtime. 00:06:36.233 [2024-11-20 15:54:34.233727] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:06:36.233 [2024-11-20 15:54:34.233737] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:06:36.233 [2024-11-20 15:54:34.233745] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid56076 for offline analysis/debug. 00:06:36.233 [2024-11-20 15:54:34.234583] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:36.806 15:54:34 rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:36.806 15:54:34 rpc -- common/autotest_common.sh@868 -- # return 0 00:06:36.807 15:54:34 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:06:36.807 15:54:34 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:06:36.807 15:54:34 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:06:36.807 15:54:34 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:06:36.807 15:54:34 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:36.807 15:54:34 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:36.807 15:54:34 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:36.807 ************************************ 00:06:36.807 START TEST rpc_integrity 00:06:36.807 ************************************ 00:06:36.807 15:54:34 rpc.rpc_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:06:36.807 15:54:34 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:06:36.807 15:54:34 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:36.807 15:54:34 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:36.807 15:54:34 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:36.807 15:54:34 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:06:36.807 15:54:34 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:06:36.807 15:54:34 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:06:36.807 15:54:34 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:06:36.807 15:54:34 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:36.807 15:54:34 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:36.807 15:54:34 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:36.807 15:54:34 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:06:36.807 15:54:34 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:06:36.807 15:54:34 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:36.807 15:54:34 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:36.807 15:54:34 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:36.807 15:54:34 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:06:36.807 { 00:06:36.807 "name": "Malloc0", 00:06:36.807 "aliases": [ 00:06:36.807 "d6ee61dd-d1d1-4627-83e0-82443f50d237" 00:06:36.807 ], 00:06:36.807 "product_name": "Malloc disk", 00:06:36.807 "block_size": 512, 00:06:36.807 "num_blocks": 16384, 00:06:36.807 "uuid": "d6ee61dd-d1d1-4627-83e0-82443f50d237", 00:06:36.807 "assigned_rate_limits": { 00:06:36.807 "rw_ios_per_sec": 0, 00:06:36.807 "rw_mbytes_per_sec": 0, 00:06:36.807 "r_mbytes_per_sec": 0, 00:06:36.807 "w_mbytes_per_sec": 0 00:06:36.807 }, 00:06:36.807 "claimed": false, 00:06:36.807 "zoned": false, 00:06:36.807 "supported_io_types": { 00:06:36.807 "read": true, 00:06:36.807 "write": true, 00:06:36.807 "unmap": true, 00:06:36.807 "flush": true, 00:06:36.807 "reset": true, 00:06:36.807 "nvme_admin": false, 00:06:36.807 "nvme_io": false, 00:06:36.807 "nvme_io_md": false, 00:06:36.807 "write_zeroes": true, 00:06:36.807 "zcopy": true, 00:06:36.807 "get_zone_info": false, 00:06:36.807 "zone_management": false, 00:06:36.807 "zone_append": false, 00:06:36.807 "compare": false, 00:06:36.807 "compare_and_write": false, 00:06:36.807 "abort": true, 00:06:36.807 "seek_hole": false, 00:06:36.807 "seek_data": false, 00:06:36.807 "copy": true, 00:06:36.807 "nvme_iov_md": false 00:06:36.807 }, 00:06:36.807 "memory_domains": [ 00:06:36.807 { 00:06:36.807 "dma_device_id": "system", 00:06:36.807 "dma_device_type": 1 00:06:36.807 }, 00:06:36.807 { 00:06:36.807 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:36.807 "dma_device_type": 2 00:06:36.807 } 00:06:36.807 ], 00:06:36.807 "driver_specific": {} 00:06:36.807 } 00:06:36.807 ]' 00:06:36.807 15:54:34 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:06:36.807 15:54:34 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:06:36.807 15:54:34 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:06:36.807 15:54:34 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:36.807 15:54:34 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:36.807 [2024-11-20 15:54:34.952031] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:06:36.807 [2024-11-20 15:54:34.952096] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:36.807 [2024-11-20 15:54:34.952127] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:06:36.807 [2024-11-20 15:54:34.952143] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:36.807 [2024-11-20 15:54:34.954515] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:36.807 [2024-11-20 15:54:34.954554] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:06:36.807 Passthru0 00:06:36.807 15:54:34 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:36.807 15:54:34 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:06:36.807 15:54:34 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:36.807 15:54:34 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:36.807 15:54:34 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:36.807 15:54:34 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:06:36.807 { 00:06:36.807 "name": "Malloc0", 00:06:36.807 "aliases": [ 00:06:36.807 "d6ee61dd-d1d1-4627-83e0-82443f50d237" 00:06:36.807 ], 00:06:36.807 "product_name": "Malloc disk", 00:06:36.807 "block_size": 512, 00:06:36.807 "num_blocks": 16384, 00:06:36.807 "uuid": "d6ee61dd-d1d1-4627-83e0-82443f50d237", 00:06:36.807 "assigned_rate_limits": { 00:06:36.807 "rw_ios_per_sec": 0, 00:06:36.807 "rw_mbytes_per_sec": 0, 00:06:36.807 "r_mbytes_per_sec": 0, 00:06:36.807 "w_mbytes_per_sec": 0 00:06:36.807 }, 00:06:36.807 "claimed": true, 00:06:36.807 "claim_type": "exclusive_write", 00:06:36.807 "zoned": false, 00:06:36.807 "supported_io_types": { 00:06:36.807 "read": true, 00:06:36.807 "write": true, 00:06:36.807 "unmap": true, 00:06:36.807 "flush": true, 00:06:36.807 "reset": true, 00:06:36.807 "nvme_admin": false, 00:06:36.807 "nvme_io": false, 00:06:36.807 "nvme_io_md": false, 00:06:36.807 "write_zeroes": true, 00:06:36.807 "zcopy": true, 00:06:36.807 "get_zone_info": false, 00:06:36.807 "zone_management": false, 00:06:36.807 "zone_append": false, 00:06:36.807 "compare": false, 00:06:36.807 "compare_and_write": false, 00:06:36.807 "abort": true, 00:06:36.807 "seek_hole": false, 00:06:36.807 "seek_data": false, 00:06:36.807 "copy": true, 00:06:36.807 "nvme_iov_md": false 00:06:36.807 }, 00:06:36.807 "memory_domains": [ 00:06:36.807 { 00:06:36.807 "dma_device_id": "system", 00:06:36.807 "dma_device_type": 1 00:06:36.807 }, 00:06:36.807 { 00:06:36.807 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:36.807 "dma_device_type": 2 00:06:36.807 } 00:06:36.807 ], 00:06:36.807 "driver_specific": {} 00:06:36.807 }, 00:06:36.807 { 00:06:36.807 "name": "Passthru0", 00:06:36.807 "aliases": [ 00:06:36.807 "932cfcee-a5e1-58b6-bdbc-105a2e2b9343" 00:06:36.807 ], 00:06:36.807 "product_name": "passthru", 00:06:36.807 "block_size": 512, 00:06:36.807 "num_blocks": 16384, 00:06:36.807 "uuid": "932cfcee-a5e1-58b6-bdbc-105a2e2b9343", 00:06:36.807 "assigned_rate_limits": { 00:06:36.807 "rw_ios_per_sec": 0, 00:06:36.807 "rw_mbytes_per_sec": 0, 00:06:36.807 "r_mbytes_per_sec": 0, 00:06:36.807 "w_mbytes_per_sec": 0 00:06:36.807 }, 00:06:36.807 "claimed": false, 00:06:36.807 "zoned": false, 00:06:36.807 "supported_io_types": { 00:06:36.807 "read": true, 00:06:36.807 "write": true, 00:06:36.807 "unmap": true, 00:06:36.807 "flush": true, 00:06:36.807 "reset": true, 00:06:36.807 "nvme_admin": false, 00:06:36.807 "nvme_io": false, 00:06:36.807 "nvme_io_md": false, 00:06:36.807 "write_zeroes": true, 00:06:36.807 "zcopy": true, 00:06:36.807 "get_zone_info": false, 00:06:36.807 "zone_management": false, 00:06:36.807 "zone_append": false, 00:06:36.807 "compare": false, 00:06:36.807 "compare_and_write": false, 00:06:36.807 "abort": true, 00:06:36.807 "seek_hole": false, 00:06:36.807 "seek_data": false, 00:06:36.807 "copy": true, 00:06:36.807 "nvme_iov_md": false 00:06:36.807 }, 00:06:36.807 "memory_domains": [ 00:06:36.807 { 00:06:36.807 "dma_device_id": "system", 00:06:36.807 "dma_device_type": 1 00:06:36.807 }, 00:06:36.807 { 00:06:36.807 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:36.807 "dma_device_type": 2 00:06:36.807 } 00:06:36.807 ], 00:06:36.807 "driver_specific": { 00:06:36.807 "passthru": { 00:06:36.807 "name": "Passthru0", 00:06:36.807 "base_bdev_name": "Malloc0" 00:06:36.807 } 00:06:36.808 } 00:06:36.808 } 00:06:36.808 ]' 00:06:36.808 15:54:34 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:06:36.808 15:54:35 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:06:36.808 15:54:35 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:06:36.808 15:54:35 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:36.808 15:54:35 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:36.808 15:54:35 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:36.808 15:54:35 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:06:36.808 15:54:35 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:36.808 15:54:35 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:36.808 15:54:35 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:36.808 15:54:35 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:06:36.808 15:54:35 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:36.808 15:54:35 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:36.808 15:54:35 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:36.808 15:54:35 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:06:36.808 15:54:35 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:06:37.068 15:54:35 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:06:37.068 00:06:37.068 real 0m0.244s 00:06:37.068 user 0m0.117s 00:06:37.068 sys 0m0.042s 00:06:37.068 15:54:35 rpc.rpc_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:37.068 15:54:35 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:37.068 ************************************ 00:06:37.068 END TEST rpc_integrity 00:06:37.068 ************************************ 00:06:37.068 15:54:35 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:06:37.068 15:54:35 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:37.068 15:54:35 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:37.068 15:54:35 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:37.068 ************************************ 00:06:37.068 START TEST rpc_plugins 00:06:37.068 ************************************ 00:06:37.068 15:54:35 rpc.rpc_plugins -- common/autotest_common.sh@1129 -- # rpc_plugins 00:06:37.068 15:54:35 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:06:37.068 15:54:35 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:37.068 15:54:35 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:37.068 15:54:35 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:37.068 15:54:35 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:06:37.068 15:54:35 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:06:37.068 15:54:35 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:37.068 15:54:35 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:37.068 15:54:35 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:37.068 15:54:35 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:06:37.068 { 00:06:37.068 "name": "Malloc1", 00:06:37.068 "aliases": [ 00:06:37.068 "fb3315a2-c26f-4422-99c0-34965fa28a27" 00:06:37.068 ], 00:06:37.068 "product_name": "Malloc disk", 00:06:37.068 "block_size": 4096, 00:06:37.068 "num_blocks": 256, 00:06:37.068 "uuid": "fb3315a2-c26f-4422-99c0-34965fa28a27", 00:06:37.068 "assigned_rate_limits": { 00:06:37.068 "rw_ios_per_sec": 0, 00:06:37.068 "rw_mbytes_per_sec": 0, 00:06:37.068 "r_mbytes_per_sec": 0, 00:06:37.068 "w_mbytes_per_sec": 0 00:06:37.068 }, 00:06:37.068 "claimed": false, 00:06:37.068 "zoned": false, 00:06:37.068 "supported_io_types": { 00:06:37.068 "read": true, 00:06:37.068 "write": true, 00:06:37.068 "unmap": true, 00:06:37.068 "flush": true, 00:06:37.068 "reset": true, 00:06:37.068 "nvme_admin": false, 00:06:37.068 "nvme_io": false, 00:06:37.068 "nvme_io_md": false, 00:06:37.068 "write_zeroes": true, 00:06:37.068 "zcopy": true, 00:06:37.068 "get_zone_info": false, 00:06:37.068 "zone_management": false, 00:06:37.068 "zone_append": false, 00:06:37.068 "compare": false, 00:06:37.068 "compare_and_write": false, 00:06:37.068 "abort": true, 00:06:37.068 "seek_hole": false, 00:06:37.068 "seek_data": false, 00:06:37.068 "copy": true, 00:06:37.068 "nvme_iov_md": false 00:06:37.068 }, 00:06:37.068 "memory_domains": [ 00:06:37.068 { 00:06:37.068 "dma_device_id": "system", 00:06:37.068 "dma_device_type": 1 00:06:37.068 }, 00:06:37.068 { 00:06:37.068 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:37.068 "dma_device_type": 2 00:06:37.068 } 00:06:37.068 ], 00:06:37.068 "driver_specific": {} 00:06:37.068 } 00:06:37.068 ]' 00:06:37.068 15:54:35 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:06:37.068 15:54:35 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:06:37.068 15:54:35 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:06:37.068 15:54:35 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:37.068 15:54:35 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:37.068 15:54:35 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:37.068 15:54:35 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:06:37.068 15:54:35 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:37.068 15:54:35 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:37.068 15:54:35 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:37.068 15:54:35 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:06:37.068 15:54:35 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:06:37.068 15:54:35 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:06:37.069 00:06:37.069 real 0m0.106s 00:06:37.069 user 0m0.059s 00:06:37.069 sys 0m0.012s 00:06:37.069 15:54:35 rpc.rpc_plugins -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:37.069 15:54:35 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:37.069 ************************************ 00:06:37.069 END TEST rpc_plugins 00:06:37.069 ************************************ 00:06:37.069 15:54:35 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:06:37.069 15:54:35 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:37.069 15:54:35 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:37.069 15:54:35 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:37.069 ************************************ 00:06:37.069 START TEST rpc_trace_cmd_test 00:06:37.069 ************************************ 00:06:37.069 15:54:35 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1129 -- # rpc_trace_cmd_test 00:06:37.069 15:54:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:06:37.069 15:54:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:06:37.069 15:54:35 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:37.069 15:54:35 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:06:37.069 15:54:35 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:37.069 15:54:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:06:37.069 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid56076", 00:06:37.069 "tpoint_group_mask": "0x8", 00:06:37.069 "iscsi_conn": { 00:06:37.069 "mask": "0x2", 00:06:37.069 "tpoint_mask": "0x0" 00:06:37.069 }, 00:06:37.069 "scsi": { 00:06:37.069 "mask": "0x4", 00:06:37.069 "tpoint_mask": "0x0" 00:06:37.069 }, 00:06:37.069 "bdev": { 00:06:37.069 "mask": "0x8", 00:06:37.069 "tpoint_mask": "0xffffffffffffffff" 00:06:37.069 }, 00:06:37.069 "nvmf_rdma": { 00:06:37.069 "mask": "0x10", 00:06:37.069 "tpoint_mask": "0x0" 00:06:37.069 }, 00:06:37.069 "nvmf_tcp": { 00:06:37.069 "mask": "0x20", 00:06:37.069 "tpoint_mask": "0x0" 00:06:37.069 }, 00:06:37.069 "ftl": { 00:06:37.069 "mask": "0x40", 00:06:37.069 "tpoint_mask": "0x0" 00:06:37.069 }, 00:06:37.069 "blobfs": { 00:06:37.069 "mask": "0x80", 00:06:37.069 "tpoint_mask": "0x0" 00:06:37.069 }, 00:06:37.069 "dsa": { 00:06:37.069 "mask": "0x200", 00:06:37.069 "tpoint_mask": "0x0" 00:06:37.069 }, 00:06:37.069 "thread": { 00:06:37.069 "mask": "0x400", 00:06:37.069 "tpoint_mask": "0x0" 00:06:37.069 }, 00:06:37.069 "nvme_pcie": { 00:06:37.069 "mask": "0x800", 00:06:37.069 "tpoint_mask": "0x0" 00:06:37.069 }, 00:06:37.069 "iaa": { 00:06:37.069 "mask": "0x1000", 00:06:37.069 "tpoint_mask": "0x0" 00:06:37.069 }, 00:06:37.069 "nvme_tcp": { 00:06:37.069 "mask": "0x2000", 00:06:37.069 "tpoint_mask": "0x0" 00:06:37.069 }, 00:06:37.069 "bdev_nvme": { 00:06:37.069 "mask": "0x4000", 00:06:37.069 "tpoint_mask": "0x0" 00:06:37.069 }, 00:06:37.069 "sock": { 00:06:37.069 "mask": "0x8000", 00:06:37.069 "tpoint_mask": "0x0" 00:06:37.069 }, 00:06:37.069 "blob": { 00:06:37.069 "mask": "0x10000", 00:06:37.069 "tpoint_mask": "0x0" 00:06:37.069 }, 00:06:37.069 "bdev_raid": { 00:06:37.069 "mask": "0x20000", 00:06:37.069 "tpoint_mask": "0x0" 00:06:37.069 }, 00:06:37.069 "scheduler": { 00:06:37.069 "mask": "0x40000", 00:06:37.069 "tpoint_mask": "0x0" 00:06:37.069 } 00:06:37.069 }' 00:06:37.069 15:54:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:06:37.327 15:54:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 19 -gt 2 ']' 00:06:37.327 15:54:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:06:37.327 15:54:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:06:37.327 15:54:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:06:37.327 15:54:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:06:37.327 15:54:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:06:37.327 15:54:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:06:37.327 15:54:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:06:37.327 15:54:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:06:37.327 00:06:37.327 real 0m0.173s 00:06:37.327 user 0m0.140s 00:06:37.327 sys 0m0.023s 00:06:37.327 15:54:35 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:37.327 ************************************ 00:06:37.327 END TEST rpc_trace_cmd_test 00:06:37.327 15:54:35 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:06:37.327 ************************************ 00:06:37.327 15:54:35 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:06:37.327 15:54:35 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:06:37.327 15:54:35 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:06:37.327 15:54:35 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:37.327 15:54:35 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:37.327 15:54:35 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:37.327 ************************************ 00:06:37.327 START TEST rpc_daemon_integrity 00:06:37.327 ************************************ 00:06:37.327 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:06:37.327 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:06:37.327 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:37.327 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:37.327 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:37.327 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:06:37.327 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:06:37.327 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:06:37.327 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:06:37.327 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:37.327 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:37.327 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:37.327 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:06:37.327 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:06:37.327 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:37.327 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:06:37.585 { 00:06:37.585 "name": "Malloc2", 00:06:37.585 "aliases": [ 00:06:37.585 "0e2ae759-ee0f-4297-9165-eb1204475c9c" 00:06:37.585 ], 00:06:37.585 "product_name": "Malloc disk", 00:06:37.585 "block_size": 512, 00:06:37.585 "num_blocks": 16384, 00:06:37.585 "uuid": "0e2ae759-ee0f-4297-9165-eb1204475c9c", 00:06:37.585 "assigned_rate_limits": { 00:06:37.585 "rw_ios_per_sec": 0, 00:06:37.585 "rw_mbytes_per_sec": 0, 00:06:37.585 "r_mbytes_per_sec": 0, 00:06:37.585 "w_mbytes_per_sec": 0 00:06:37.585 }, 00:06:37.585 "claimed": false, 00:06:37.585 "zoned": false, 00:06:37.585 "supported_io_types": { 00:06:37.585 "read": true, 00:06:37.585 "write": true, 00:06:37.585 "unmap": true, 00:06:37.585 "flush": true, 00:06:37.585 "reset": true, 00:06:37.585 "nvme_admin": false, 00:06:37.585 "nvme_io": false, 00:06:37.585 "nvme_io_md": false, 00:06:37.585 "write_zeroes": true, 00:06:37.585 "zcopy": true, 00:06:37.585 "get_zone_info": false, 00:06:37.585 "zone_management": false, 00:06:37.585 "zone_append": false, 00:06:37.585 "compare": false, 00:06:37.585 "compare_and_write": false, 00:06:37.585 "abort": true, 00:06:37.585 "seek_hole": false, 00:06:37.585 "seek_data": false, 00:06:37.585 "copy": true, 00:06:37.585 "nvme_iov_md": false 00:06:37.585 }, 00:06:37.585 "memory_domains": [ 00:06:37.585 { 00:06:37.585 "dma_device_id": "system", 00:06:37.585 "dma_device_type": 1 00:06:37.585 }, 00:06:37.585 { 00:06:37.585 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:37.585 "dma_device_type": 2 00:06:37.585 } 00:06:37.585 ], 00:06:37.585 "driver_specific": {} 00:06:37.585 } 00:06:37.585 ]' 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:37.585 [2024-11-20 15:54:35.612017] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:06:37.585 [2024-11-20 15:54:35.612079] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:37.585 [2024-11-20 15:54:35.612100] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:06:37.585 [2024-11-20 15:54:35.612111] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:37.585 [2024-11-20 15:54:35.614295] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:37.585 [2024-11-20 15:54:35.614333] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:06:37.585 Passthru0 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:06:37.585 { 00:06:37.585 "name": "Malloc2", 00:06:37.585 "aliases": [ 00:06:37.585 "0e2ae759-ee0f-4297-9165-eb1204475c9c" 00:06:37.585 ], 00:06:37.585 "product_name": "Malloc disk", 00:06:37.585 "block_size": 512, 00:06:37.585 "num_blocks": 16384, 00:06:37.585 "uuid": "0e2ae759-ee0f-4297-9165-eb1204475c9c", 00:06:37.585 "assigned_rate_limits": { 00:06:37.585 "rw_ios_per_sec": 0, 00:06:37.585 "rw_mbytes_per_sec": 0, 00:06:37.585 "r_mbytes_per_sec": 0, 00:06:37.585 "w_mbytes_per_sec": 0 00:06:37.585 }, 00:06:37.585 "claimed": true, 00:06:37.585 "claim_type": "exclusive_write", 00:06:37.585 "zoned": false, 00:06:37.585 "supported_io_types": { 00:06:37.585 "read": true, 00:06:37.585 "write": true, 00:06:37.585 "unmap": true, 00:06:37.585 "flush": true, 00:06:37.585 "reset": true, 00:06:37.585 "nvme_admin": false, 00:06:37.585 "nvme_io": false, 00:06:37.585 "nvme_io_md": false, 00:06:37.585 "write_zeroes": true, 00:06:37.585 "zcopy": true, 00:06:37.585 "get_zone_info": false, 00:06:37.585 "zone_management": false, 00:06:37.585 "zone_append": false, 00:06:37.585 "compare": false, 00:06:37.585 "compare_and_write": false, 00:06:37.585 "abort": true, 00:06:37.585 "seek_hole": false, 00:06:37.585 "seek_data": false, 00:06:37.585 "copy": true, 00:06:37.585 "nvme_iov_md": false 00:06:37.585 }, 00:06:37.585 "memory_domains": [ 00:06:37.585 { 00:06:37.585 "dma_device_id": "system", 00:06:37.585 "dma_device_type": 1 00:06:37.585 }, 00:06:37.585 { 00:06:37.585 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:37.585 "dma_device_type": 2 00:06:37.585 } 00:06:37.585 ], 00:06:37.585 "driver_specific": {} 00:06:37.585 }, 00:06:37.585 { 00:06:37.585 "name": "Passthru0", 00:06:37.585 "aliases": [ 00:06:37.585 "6016c6ce-c189-537f-9813-f0e54eb8972b" 00:06:37.585 ], 00:06:37.585 "product_name": "passthru", 00:06:37.585 "block_size": 512, 00:06:37.585 "num_blocks": 16384, 00:06:37.585 "uuid": "6016c6ce-c189-537f-9813-f0e54eb8972b", 00:06:37.585 "assigned_rate_limits": { 00:06:37.585 "rw_ios_per_sec": 0, 00:06:37.585 "rw_mbytes_per_sec": 0, 00:06:37.585 "r_mbytes_per_sec": 0, 00:06:37.585 "w_mbytes_per_sec": 0 00:06:37.585 }, 00:06:37.585 "claimed": false, 00:06:37.585 "zoned": false, 00:06:37.585 "supported_io_types": { 00:06:37.585 "read": true, 00:06:37.585 "write": true, 00:06:37.585 "unmap": true, 00:06:37.585 "flush": true, 00:06:37.585 "reset": true, 00:06:37.585 "nvme_admin": false, 00:06:37.585 "nvme_io": false, 00:06:37.585 "nvme_io_md": false, 00:06:37.585 "write_zeroes": true, 00:06:37.585 "zcopy": true, 00:06:37.585 "get_zone_info": false, 00:06:37.585 "zone_management": false, 00:06:37.585 "zone_append": false, 00:06:37.585 "compare": false, 00:06:37.585 "compare_and_write": false, 00:06:37.585 "abort": true, 00:06:37.585 "seek_hole": false, 00:06:37.585 "seek_data": false, 00:06:37.585 "copy": true, 00:06:37.585 "nvme_iov_md": false 00:06:37.585 }, 00:06:37.585 "memory_domains": [ 00:06:37.585 { 00:06:37.585 "dma_device_id": "system", 00:06:37.585 "dma_device_type": 1 00:06:37.585 }, 00:06:37.585 { 00:06:37.585 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:37.585 "dma_device_type": 2 00:06:37.585 } 00:06:37.585 ], 00:06:37.585 "driver_specific": { 00:06:37.585 "passthru": { 00:06:37.585 "name": "Passthru0", 00:06:37.585 "base_bdev_name": "Malloc2" 00:06:37.585 } 00:06:37.585 } 00:06:37.585 } 00:06:37.585 ]' 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:37.585 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:06:37.586 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:37.586 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:37.586 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:37.586 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:06:37.586 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:37.586 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:37.586 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:37.586 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:06:37.586 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:06:37.586 15:54:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:06:37.586 00:06:37.586 real 0m0.247s 00:06:37.586 user 0m0.140s 00:06:37.586 sys 0m0.025s 00:06:37.586 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:37.586 15:54:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:37.586 ************************************ 00:06:37.586 END TEST rpc_daemon_integrity 00:06:37.586 ************************************ 00:06:37.586 15:54:35 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:06:37.586 15:54:35 rpc -- rpc/rpc.sh@84 -- # killprocess 56076 00:06:37.586 15:54:35 rpc -- common/autotest_common.sh@954 -- # '[' -z 56076 ']' 00:06:37.586 15:54:35 rpc -- common/autotest_common.sh@958 -- # kill -0 56076 00:06:37.586 15:54:35 rpc -- common/autotest_common.sh@959 -- # uname 00:06:37.586 15:54:35 rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:37.586 15:54:35 rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 56076 00:06:37.586 15:54:35 rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:37.586 15:54:35 rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:37.586 15:54:35 rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 56076' 00:06:37.586 killing process with pid 56076 00:06:37.586 15:54:35 rpc -- common/autotest_common.sh@973 -- # kill 56076 00:06:37.586 15:54:35 rpc -- common/autotest_common.sh@978 -- # wait 56076 00:06:39.566 00:06:39.566 real 0m3.592s 00:06:39.566 user 0m4.026s 00:06:39.566 sys 0m0.578s 00:06:39.566 ************************************ 00:06:39.566 15:54:37 rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:39.566 15:54:37 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:39.566 END TEST rpc 00:06:39.566 ************************************ 00:06:39.566 15:54:37 -- spdk/autotest.sh@157 -- # run_test skip_rpc /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:06:39.566 15:54:37 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:39.566 15:54:37 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:39.566 15:54:37 -- common/autotest_common.sh@10 -- # set +x 00:06:39.566 ************************************ 00:06:39.566 START TEST skip_rpc 00:06:39.566 ************************************ 00:06:39.566 15:54:37 skip_rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:06:39.566 * Looking for test storage... 00:06:39.566 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:06:39.566 15:54:37 skip_rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:39.566 15:54:37 skip_rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:06:39.566 15:54:37 skip_rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:39.566 15:54:37 skip_rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@344 -- # case "$op" in 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@345 -- # : 1 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@365 -- # decimal 1 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@353 -- # local d=1 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@355 -- # echo 1 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@366 -- # decimal 2 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@353 -- # local d=2 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@355 -- # echo 2 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:39.566 15:54:37 skip_rpc -- scripts/common.sh@368 -- # return 0 00:06:39.566 15:54:37 skip_rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:39.566 15:54:37 skip_rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:39.566 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:39.566 --rc genhtml_branch_coverage=1 00:06:39.566 --rc genhtml_function_coverage=1 00:06:39.566 --rc genhtml_legend=1 00:06:39.566 --rc geninfo_all_blocks=1 00:06:39.566 --rc geninfo_unexecuted_blocks=1 00:06:39.566 00:06:39.566 ' 00:06:39.566 15:54:37 skip_rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:39.566 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:39.566 --rc genhtml_branch_coverage=1 00:06:39.566 --rc genhtml_function_coverage=1 00:06:39.566 --rc genhtml_legend=1 00:06:39.566 --rc geninfo_all_blocks=1 00:06:39.566 --rc geninfo_unexecuted_blocks=1 00:06:39.566 00:06:39.566 ' 00:06:39.566 15:54:37 skip_rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:39.566 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:39.566 --rc genhtml_branch_coverage=1 00:06:39.566 --rc genhtml_function_coverage=1 00:06:39.566 --rc genhtml_legend=1 00:06:39.566 --rc geninfo_all_blocks=1 00:06:39.566 --rc geninfo_unexecuted_blocks=1 00:06:39.566 00:06:39.566 ' 00:06:39.566 15:54:37 skip_rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:39.566 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:39.566 --rc genhtml_branch_coverage=1 00:06:39.566 --rc genhtml_function_coverage=1 00:06:39.566 --rc genhtml_legend=1 00:06:39.566 --rc geninfo_all_blocks=1 00:06:39.566 --rc geninfo_unexecuted_blocks=1 00:06:39.566 00:06:39.566 ' 00:06:39.566 15:54:37 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:06:39.566 15:54:37 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:06:39.566 15:54:37 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:06:39.566 15:54:37 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:39.566 15:54:37 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:39.566 15:54:37 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:39.566 ************************************ 00:06:39.566 START TEST skip_rpc 00:06:39.566 ************************************ 00:06:39.566 15:54:37 skip_rpc.skip_rpc -- common/autotest_common.sh@1129 -- # test_skip_rpc 00:06:39.566 15:54:37 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=56293 00:06:39.566 15:54:37 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:39.566 15:54:37 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:06:39.566 15:54:37 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:06:39.566 [2024-11-20 15:54:37.640924] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:06:39.566 [2024-11-20 15:54:37.641056] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid56293 ] 00:06:39.566 [2024-11-20 15:54:37.793017] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:39.826 [2024-11-20 15:54:37.892493] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@652 -- # local es=0 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd spdk_get_version 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # rpc_cmd spdk_get_version 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # es=1 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 56293 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # '[' -z 56293 ']' 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@958 -- # kill -0 56293 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # uname 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 56293 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:45.112 killing process with pid 56293 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 56293' 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@973 -- # kill 56293 00:06:45.112 15:54:42 skip_rpc.skip_rpc -- common/autotest_common.sh@978 -- # wait 56293 00:06:46.054 00:06:46.054 real 0m6.551s 00:06:46.054 user 0m6.174s 00:06:46.054 sys 0m0.273s 00:06:46.054 15:54:44 skip_rpc.skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:46.054 15:54:44 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:46.054 ************************************ 00:06:46.054 END TEST skip_rpc 00:06:46.054 ************************************ 00:06:46.054 15:54:44 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:06:46.054 15:54:44 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:46.054 15:54:44 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:46.054 15:54:44 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:46.054 ************************************ 00:06:46.054 START TEST skip_rpc_with_json 00:06:46.054 ************************************ 00:06:46.054 15:54:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_json 00:06:46.054 15:54:44 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:06:46.054 15:54:44 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=56392 00:06:46.054 15:54:44 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:46.054 15:54:44 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 56392 00:06:46.054 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:46.054 15:54:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@835 -- # '[' -z 56392 ']' 00:06:46.054 15:54:44 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:46.054 15:54:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:46.054 15:54:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:46.054 15:54:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:46.054 15:54:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:46.054 15:54:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:46.054 [2024-11-20 15:54:44.247879] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:06:46.054 [2024-11-20 15:54:44.248012] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid56392 ] 00:06:46.315 [2024-11-20 15:54:44.406115] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:46.315 [2024-11-20 15:54:44.508860] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:46.886 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:46.886 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@868 -- # return 0 00:06:46.887 15:54:45 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:06:46.887 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:46.887 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:47.147 [2024-11-20 15:54:45.139409] nvmf_rpc.c:2706:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:06:47.147 request: 00:06:47.147 { 00:06:47.147 "trtype": "tcp", 00:06:47.147 "method": "nvmf_get_transports", 00:06:47.147 "req_id": 1 00:06:47.147 } 00:06:47.147 Got JSON-RPC error response 00:06:47.147 response: 00:06:47.147 { 00:06:47.147 "code": -19, 00:06:47.147 "message": "No such device" 00:06:47.147 } 00:06:47.147 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:06:47.147 15:54:45 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:06:47.147 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.147 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:47.147 [2024-11-20 15:54:45.151495] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:06:47.147 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.147 15:54:45 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:06:47.147 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.147 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:47.147 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:47.147 15:54:45 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:06:47.147 { 00:06:47.147 "subsystems": [ 00:06:47.147 { 00:06:47.147 "subsystem": "fsdev", 00:06:47.147 "config": [ 00:06:47.147 { 00:06:47.147 "method": "fsdev_set_opts", 00:06:47.147 "params": { 00:06:47.147 "fsdev_io_pool_size": 65535, 00:06:47.147 "fsdev_io_cache_size": 256 00:06:47.147 } 00:06:47.147 } 00:06:47.147 ] 00:06:47.147 }, 00:06:47.147 { 00:06:47.147 "subsystem": "keyring", 00:06:47.147 "config": [] 00:06:47.147 }, 00:06:47.147 { 00:06:47.147 "subsystem": "iobuf", 00:06:47.147 "config": [ 00:06:47.147 { 00:06:47.147 "method": "iobuf_set_options", 00:06:47.147 "params": { 00:06:47.147 "small_pool_count": 8192, 00:06:47.147 "large_pool_count": 1024, 00:06:47.147 "small_bufsize": 8192, 00:06:47.147 "large_bufsize": 135168, 00:06:47.147 "enable_numa": false 00:06:47.147 } 00:06:47.147 } 00:06:47.147 ] 00:06:47.147 }, 00:06:47.147 { 00:06:47.147 "subsystem": "sock", 00:06:47.147 "config": [ 00:06:47.147 { 00:06:47.147 "method": "sock_set_default_impl", 00:06:47.147 "params": { 00:06:47.147 "impl_name": "posix" 00:06:47.147 } 00:06:47.147 }, 00:06:47.147 { 00:06:47.147 "method": "sock_impl_set_options", 00:06:47.147 "params": { 00:06:47.147 "impl_name": "ssl", 00:06:47.147 "recv_buf_size": 4096, 00:06:47.147 "send_buf_size": 4096, 00:06:47.147 "enable_recv_pipe": true, 00:06:47.147 "enable_quickack": false, 00:06:47.147 "enable_placement_id": 0, 00:06:47.147 "enable_zerocopy_send_server": true, 00:06:47.147 "enable_zerocopy_send_client": false, 00:06:47.147 "zerocopy_threshold": 0, 00:06:47.147 "tls_version": 0, 00:06:47.147 "enable_ktls": false 00:06:47.147 } 00:06:47.147 }, 00:06:47.147 { 00:06:47.147 "method": "sock_impl_set_options", 00:06:47.147 "params": { 00:06:47.147 "impl_name": "posix", 00:06:47.147 "recv_buf_size": 2097152, 00:06:47.147 "send_buf_size": 2097152, 00:06:47.147 "enable_recv_pipe": true, 00:06:47.147 "enable_quickack": false, 00:06:47.147 "enable_placement_id": 0, 00:06:47.147 "enable_zerocopy_send_server": true, 00:06:47.147 "enable_zerocopy_send_client": false, 00:06:47.147 "zerocopy_threshold": 0, 00:06:47.147 "tls_version": 0, 00:06:47.147 "enable_ktls": false 00:06:47.147 } 00:06:47.147 } 00:06:47.147 ] 00:06:47.147 }, 00:06:47.147 { 00:06:47.147 "subsystem": "vmd", 00:06:47.147 "config": [] 00:06:47.147 }, 00:06:47.147 { 00:06:47.147 "subsystem": "accel", 00:06:47.147 "config": [ 00:06:47.147 { 00:06:47.147 "method": "accel_set_options", 00:06:47.147 "params": { 00:06:47.147 "small_cache_size": 128, 00:06:47.147 "large_cache_size": 16, 00:06:47.147 "task_count": 2048, 00:06:47.147 "sequence_count": 2048, 00:06:47.147 "buf_count": 2048 00:06:47.147 } 00:06:47.147 } 00:06:47.147 ] 00:06:47.147 }, 00:06:47.147 { 00:06:47.147 "subsystem": "bdev", 00:06:47.147 "config": [ 00:06:47.147 { 00:06:47.147 "method": "bdev_set_options", 00:06:47.147 "params": { 00:06:47.147 "bdev_io_pool_size": 65535, 00:06:47.147 "bdev_io_cache_size": 256, 00:06:47.147 "bdev_auto_examine": true, 00:06:47.147 "iobuf_small_cache_size": 128, 00:06:47.147 "iobuf_large_cache_size": 16 00:06:47.147 } 00:06:47.147 }, 00:06:47.147 { 00:06:47.147 "method": "bdev_raid_set_options", 00:06:47.147 "params": { 00:06:47.147 "process_window_size_kb": 1024, 00:06:47.147 "process_max_bandwidth_mb_sec": 0 00:06:47.147 } 00:06:47.147 }, 00:06:47.147 { 00:06:47.147 "method": "bdev_iscsi_set_options", 00:06:47.147 "params": { 00:06:47.147 "timeout_sec": 30 00:06:47.147 } 00:06:47.147 }, 00:06:47.147 { 00:06:47.147 "method": "bdev_nvme_set_options", 00:06:47.147 "params": { 00:06:47.147 "action_on_timeout": "none", 00:06:47.147 "timeout_us": 0, 00:06:47.147 "timeout_admin_us": 0, 00:06:47.147 "keep_alive_timeout_ms": 10000, 00:06:47.147 "arbitration_burst": 0, 00:06:47.148 "low_priority_weight": 0, 00:06:47.148 "medium_priority_weight": 0, 00:06:47.148 "high_priority_weight": 0, 00:06:47.148 "nvme_adminq_poll_period_us": 10000, 00:06:47.148 "nvme_ioq_poll_period_us": 0, 00:06:47.148 "io_queue_requests": 0, 00:06:47.148 "delay_cmd_submit": true, 00:06:47.148 "transport_retry_count": 4, 00:06:47.148 "bdev_retry_count": 3, 00:06:47.148 "transport_ack_timeout": 0, 00:06:47.148 "ctrlr_loss_timeout_sec": 0, 00:06:47.148 "reconnect_delay_sec": 0, 00:06:47.148 "fast_io_fail_timeout_sec": 0, 00:06:47.148 "disable_auto_failback": false, 00:06:47.148 "generate_uuids": false, 00:06:47.148 "transport_tos": 0, 00:06:47.148 "nvme_error_stat": false, 00:06:47.148 "rdma_srq_size": 0, 00:06:47.148 "io_path_stat": false, 00:06:47.148 "allow_accel_sequence": false, 00:06:47.148 "rdma_max_cq_size": 0, 00:06:47.148 "rdma_cm_event_timeout_ms": 0, 00:06:47.148 "dhchap_digests": [ 00:06:47.148 "sha256", 00:06:47.148 "sha384", 00:06:47.148 "sha512" 00:06:47.148 ], 00:06:47.148 "dhchap_dhgroups": [ 00:06:47.148 "null", 00:06:47.148 "ffdhe2048", 00:06:47.148 "ffdhe3072", 00:06:47.148 "ffdhe4096", 00:06:47.148 "ffdhe6144", 00:06:47.148 "ffdhe8192" 00:06:47.148 ] 00:06:47.148 } 00:06:47.148 }, 00:06:47.148 { 00:06:47.148 "method": "bdev_nvme_set_hotplug", 00:06:47.148 "params": { 00:06:47.148 "period_us": 100000, 00:06:47.148 "enable": false 00:06:47.148 } 00:06:47.148 }, 00:06:47.148 { 00:06:47.148 "method": "bdev_wait_for_examine" 00:06:47.148 } 00:06:47.148 ] 00:06:47.148 }, 00:06:47.148 { 00:06:47.148 "subsystem": "scsi", 00:06:47.148 "config": null 00:06:47.148 }, 00:06:47.148 { 00:06:47.148 "subsystem": "scheduler", 00:06:47.148 "config": [ 00:06:47.148 { 00:06:47.148 "method": "framework_set_scheduler", 00:06:47.148 "params": { 00:06:47.148 "name": "static" 00:06:47.148 } 00:06:47.148 } 00:06:47.148 ] 00:06:47.148 }, 00:06:47.148 { 00:06:47.148 "subsystem": "vhost_scsi", 00:06:47.148 "config": [] 00:06:47.148 }, 00:06:47.148 { 00:06:47.148 "subsystem": "vhost_blk", 00:06:47.148 "config": [] 00:06:47.148 }, 00:06:47.148 { 00:06:47.148 "subsystem": "ublk", 00:06:47.148 "config": [] 00:06:47.148 }, 00:06:47.148 { 00:06:47.148 "subsystem": "nbd", 00:06:47.148 "config": [] 00:06:47.148 }, 00:06:47.148 { 00:06:47.148 "subsystem": "nvmf", 00:06:47.148 "config": [ 00:06:47.148 { 00:06:47.148 "method": "nvmf_set_config", 00:06:47.148 "params": { 00:06:47.148 "discovery_filter": "match_any", 00:06:47.148 "admin_cmd_passthru": { 00:06:47.148 "identify_ctrlr": false 00:06:47.148 }, 00:06:47.148 "dhchap_digests": [ 00:06:47.148 "sha256", 00:06:47.148 "sha384", 00:06:47.148 "sha512" 00:06:47.148 ], 00:06:47.148 "dhchap_dhgroups": [ 00:06:47.148 "null", 00:06:47.148 "ffdhe2048", 00:06:47.148 "ffdhe3072", 00:06:47.148 "ffdhe4096", 00:06:47.148 "ffdhe6144", 00:06:47.148 "ffdhe8192" 00:06:47.148 ] 00:06:47.148 } 00:06:47.148 }, 00:06:47.148 { 00:06:47.148 "method": "nvmf_set_max_subsystems", 00:06:47.148 "params": { 00:06:47.148 "max_subsystems": 1024 00:06:47.148 } 00:06:47.148 }, 00:06:47.148 { 00:06:47.148 "method": "nvmf_set_crdt", 00:06:47.148 "params": { 00:06:47.148 "crdt1": 0, 00:06:47.148 "crdt2": 0, 00:06:47.148 "crdt3": 0 00:06:47.148 } 00:06:47.148 }, 00:06:47.148 { 00:06:47.148 "method": "nvmf_create_transport", 00:06:47.148 "params": { 00:06:47.148 "trtype": "TCP", 00:06:47.148 "max_queue_depth": 128, 00:06:47.148 "max_io_qpairs_per_ctrlr": 127, 00:06:47.148 "in_capsule_data_size": 4096, 00:06:47.148 "max_io_size": 131072, 00:06:47.148 "io_unit_size": 131072, 00:06:47.148 "max_aq_depth": 128, 00:06:47.148 "num_shared_buffers": 511, 00:06:47.148 "buf_cache_size": 4294967295, 00:06:47.148 "dif_insert_or_strip": false, 00:06:47.148 "zcopy": false, 00:06:47.148 "c2h_success": true, 00:06:47.148 "sock_priority": 0, 00:06:47.148 "abort_timeout_sec": 1, 00:06:47.148 "ack_timeout": 0, 00:06:47.148 "data_wr_pool_size": 0 00:06:47.148 } 00:06:47.148 } 00:06:47.148 ] 00:06:47.148 }, 00:06:47.148 { 00:06:47.148 "subsystem": "iscsi", 00:06:47.148 "config": [ 00:06:47.148 { 00:06:47.148 "method": "iscsi_set_options", 00:06:47.148 "params": { 00:06:47.148 "node_base": "iqn.2016-06.io.spdk", 00:06:47.148 "max_sessions": 128, 00:06:47.148 "max_connections_per_session": 2, 00:06:47.148 "max_queue_depth": 64, 00:06:47.148 "default_time2wait": 2, 00:06:47.148 "default_time2retain": 20, 00:06:47.148 "first_burst_length": 8192, 00:06:47.148 "immediate_data": true, 00:06:47.148 "allow_duplicated_isid": false, 00:06:47.148 "error_recovery_level": 0, 00:06:47.148 "nop_timeout": 60, 00:06:47.148 "nop_in_interval": 30, 00:06:47.148 "disable_chap": false, 00:06:47.148 "require_chap": false, 00:06:47.148 "mutual_chap": false, 00:06:47.148 "chap_group": 0, 00:06:47.148 "max_large_datain_per_connection": 64, 00:06:47.148 "max_r2t_per_connection": 4, 00:06:47.148 "pdu_pool_size": 36864, 00:06:47.148 "immediate_data_pool_size": 16384, 00:06:47.148 "data_out_pool_size": 2048 00:06:47.148 } 00:06:47.148 } 00:06:47.148 ] 00:06:47.148 } 00:06:47.148 ] 00:06:47.148 } 00:06:47.148 15:54:45 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:06:47.148 15:54:45 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 56392 00:06:47.148 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 56392 ']' 00:06:47.148 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 56392 00:06:47.148 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:06:47.148 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:47.148 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 56392 00:06:47.148 killing process with pid 56392 00:06:47.148 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:47.148 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:47.148 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 56392' 00:06:47.148 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 56392 00:06:47.148 15:54:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 56392 00:06:49.061 15:54:46 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:06:49.061 15:54:46 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=56431 00:06:49.061 15:54:46 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:06:54.348 15:54:51 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 56431 00:06:54.348 15:54:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 56431 ']' 00:06:54.348 15:54:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 56431 00:06:54.348 15:54:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:06:54.348 15:54:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:54.348 15:54:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 56431 00:06:54.348 15:54:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:54.348 15:54:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:54.348 killing process with pid 56431 00:06:54.348 15:54:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 56431' 00:06:54.348 15:54:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 56431 00:06:54.348 15:54:51 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 56431 00:06:55.287 15:54:53 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:06:55.287 15:54:53 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:06:55.287 00:06:55.287 real 0m9.206s 00:06:55.287 user 0m8.795s 00:06:55.287 sys 0m0.624s 00:06:55.287 15:54:53 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:55.287 15:54:53 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:55.287 ************************************ 00:06:55.287 END TEST skip_rpc_with_json 00:06:55.287 ************************************ 00:06:55.287 15:54:53 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:06:55.287 15:54:53 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:55.287 15:54:53 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:55.287 15:54:53 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:55.287 ************************************ 00:06:55.287 START TEST skip_rpc_with_delay 00:06:55.287 ************************************ 00:06:55.287 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_delay 00:06:55.287 15:54:53 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:55.287 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@652 -- # local es=0 00:06:55.287 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:55.287 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:55.287 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:55.287 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:55.287 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:55.287 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:55.287 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:55.287 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:55.287 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:06:55.287 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:55.287 [2024-11-20 15:54:53.509221] app.c: 842:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:06:55.548 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # es=1 00:06:55.548 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:55.548 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:55.548 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:55.548 00:06:55.548 real 0m0.124s 00:06:55.548 user 0m0.065s 00:06:55.548 sys 0m0.057s 00:06:55.548 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:55.548 15:54:53 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:06:55.548 ************************************ 00:06:55.548 END TEST skip_rpc_with_delay 00:06:55.548 ************************************ 00:06:55.548 15:54:53 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:06:55.548 15:54:53 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:06:55.548 15:54:53 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:06:55.548 15:54:53 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:55.548 15:54:53 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:55.548 15:54:53 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:55.548 ************************************ 00:06:55.548 START TEST exit_on_failed_rpc_init 00:06:55.548 ************************************ 00:06:55.548 15:54:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1129 -- # test_exit_on_failed_rpc_init 00:06:55.548 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:55.548 15:54:53 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=56554 00:06:55.548 15:54:53 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:55.548 15:54:53 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 56554 00:06:55.548 15:54:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@835 -- # '[' -z 56554 ']' 00:06:55.548 15:54:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:55.548 15:54:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:55.548 15:54:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:55.548 15:54:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:55.548 15:54:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:06:55.548 [2024-11-20 15:54:53.674751] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:06:55.548 [2024-11-20 15:54:53.674842] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid56554 ] 00:06:55.806 [2024-11-20 15:54:53.831085] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:55.807 [2024-11-20 15:54:53.931018] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:56.372 15:54:54 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:56.372 15:54:54 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@868 -- # return 0 00:06:56.372 15:54:54 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:56.372 15:54:54 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:56.372 15:54:54 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@652 -- # local es=0 00:06:56.372 15:54:54 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:56.372 15:54:54 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:56.372 15:54:54 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:56.372 15:54:54 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:56.372 15:54:54 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:56.372 15:54:54 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:56.372 15:54:54 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:56.372 15:54:54 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:56.372 15:54:54 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:06:56.372 15:54:54 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:56.372 [2024-11-20 15:54:54.596487] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:06:56.372 [2024-11-20 15:54:54.596580] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid56572 ] 00:06:56.630 [2024-11-20 15:54:54.751362] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:56.630 [2024-11-20 15:54:54.850991] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:56.630 [2024-11-20 15:54:54.851075] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:06:56.630 [2024-11-20 15:54:54.851088] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:06:56.630 [2024-11-20 15:54:54.851098] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # es=234 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@664 -- # es=106 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@665 -- # case "$es" in 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@672 -- # es=1 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 56554 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # '[' -z 56554 ']' 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@958 -- # kill -0 56554 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # uname 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 56554 00:06:56.887 killing process with pid 56554 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@972 -- # echo 'killing process with pid 56554' 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@973 -- # kill 56554 00:06:56.887 15:54:55 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@978 -- # wait 56554 00:06:58.786 ************************************ 00:06:58.786 END TEST exit_on_failed_rpc_init 00:06:58.786 ************************************ 00:06:58.786 00:06:58.786 real 0m2.957s 00:06:58.786 user 0m3.289s 00:06:58.786 sys 0m0.379s 00:06:58.786 15:54:56 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:58.786 15:54:56 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:06:58.786 15:54:56 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:06:58.786 00:06:58.786 real 0m19.202s 00:06:58.786 user 0m18.456s 00:06:58.786 sys 0m1.505s 00:06:58.786 ************************************ 00:06:58.786 END TEST skip_rpc 00:06:58.786 ************************************ 00:06:58.786 15:54:56 skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:58.786 15:54:56 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:58.786 15:54:56 -- spdk/autotest.sh@158 -- # run_test rpc_client /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:06:58.786 15:54:56 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:58.786 15:54:56 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:58.786 15:54:56 -- common/autotest_common.sh@10 -- # set +x 00:06:58.786 ************************************ 00:06:58.786 START TEST rpc_client 00:06:58.786 ************************************ 00:06:58.786 15:54:56 rpc_client -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:06:58.786 * Looking for test storage... 00:06:58.786 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc_client 00:06:58.786 15:54:56 rpc_client -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:58.786 15:54:56 rpc_client -- common/autotest_common.sh@1693 -- # lcov --version 00:06:58.786 15:54:56 rpc_client -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:58.786 15:54:56 rpc_client -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@336 -- # IFS=.-: 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@336 -- # read -ra ver1 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@337 -- # IFS=.-: 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@337 -- # read -ra ver2 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@338 -- # local 'op=<' 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@340 -- # ver1_l=2 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@341 -- # ver2_l=1 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@344 -- # case "$op" in 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@345 -- # : 1 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@365 -- # decimal 1 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@353 -- # local d=1 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@355 -- # echo 1 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@365 -- # ver1[v]=1 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@366 -- # decimal 2 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@353 -- # local d=2 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@355 -- # echo 2 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@366 -- # ver2[v]=2 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:58.786 15:54:56 rpc_client -- scripts/common.sh@368 -- # return 0 00:06:58.786 15:54:56 rpc_client -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:58.786 15:54:56 rpc_client -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:58.786 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.786 --rc genhtml_branch_coverage=1 00:06:58.786 --rc genhtml_function_coverage=1 00:06:58.786 --rc genhtml_legend=1 00:06:58.786 --rc geninfo_all_blocks=1 00:06:58.786 --rc geninfo_unexecuted_blocks=1 00:06:58.786 00:06:58.786 ' 00:06:58.786 15:54:56 rpc_client -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:58.786 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.786 --rc genhtml_branch_coverage=1 00:06:58.786 --rc genhtml_function_coverage=1 00:06:58.786 --rc genhtml_legend=1 00:06:58.786 --rc geninfo_all_blocks=1 00:06:58.786 --rc geninfo_unexecuted_blocks=1 00:06:58.786 00:06:58.786 ' 00:06:58.786 15:54:56 rpc_client -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:58.786 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.786 --rc genhtml_branch_coverage=1 00:06:58.786 --rc genhtml_function_coverage=1 00:06:58.786 --rc genhtml_legend=1 00:06:58.786 --rc geninfo_all_blocks=1 00:06:58.786 --rc geninfo_unexecuted_blocks=1 00:06:58.786 00:06:58.786 ' 00:06:58.786 15:54:56 rpc_client -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:58.786 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.786 --rc genhtml_branch_coverage=1 00:06:58.786 --rc genhtml_function_coverage=1 00:06:58.786 --rc genhtml_legend=1 00:06:58.786 --rc geninfo_all_blocks=1 00:06:58.786 --rc geninfo_unexecuted_blocks=1 00:06:58.786 00:06:58.786 ' 00:06:58.786 15:54:56 rpc_client -- rpc_client/rpc_client.sh@10 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client_test 00:06:58.786 OK 00:06:58.786 15:54:56 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:06:58.786 00:06:58.786 real 0m0.184s 00:06:58.786 user 0m0.099s 00:06:58.786 sys 0m0.090s 00:06:58.786 ************************************ 00:06:58.786 END TEST rpc_client 00:06:58.786 ************************************ 00:06:58.786 15:54:56 rpc_client -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:58.786 15:54:56 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:06:58.786 15:54:56 -- spdk/autotest.sh@159 -- # run_test json_config /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:06:58.786 15:54:56 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:58.786 15:54:56 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:58.786 15:54:56 -- common/autotest_common.sh@10 -- # set +x 00:06:58.786 ************************************ 00:06:58.786 START TEST json_config 00:06:58.786 ************************************ 00:06:58.786 15:54:56 json_config -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:06:58.786 15:54:56 json_config -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:58.786 15:54:56 json_config -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:58.786 15:54:56 json_config -- common/autotest_common.sh@1693 -- # lcov --version 00:06:58.786 15:54:56 json_config -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:58.786 15:54:56 json_config -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:58.786 15:54:56 json_config -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:58.786 15:54:56 json_config -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:58.786 15:54:56 json_config -- scripts/common.sh@336 -- # IFS=.-: 00:06:58.786 15:54:56 json_config -- scripts/common.sh@336 -- # read -ra ver1 00:06:58.786 15:54:56 json_config -- scripts/common.sh@337 -- # IFS=.-: 00:06:58.786 15:54:56 json_config -- scripts/common.sh@337 -- # read -ra ver2 00:06:58.786 15:54:56 json_config -- scripts/common.sh@338 -- # local 'op=<' 00:06:58.786 15:54:56 json_config -- scripts/common.sh@340 -- # ver1_l=2 00:06:58.786 15:54:56 json_config -- scripts/common.sh@341 -- # ver2_l=1 00:06:58.786 15:54:56 json_config -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:58.786 15:54:56 json_config -- scripts/common.sh@344 -- # case "$op" in 00:06:58.786 15:54:56 json_config -- scripts/common.sh@345 -- # : 1 00:06:58.786 15:54:56 json_config -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:58.786 15:54:56 json_config -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:58.786 15:54:56 json_config -- scripts/common.sh@365 -- # decimal 1 00:06:58.786 15:54:56 json_config -- scripts/common.sh@353 -- # local d=1 00:06:58.786 15:54:56 json_config -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:58.786 15:54:56 json_config -- scripts/common.sh@355 -- # echo 1 00:06:58.786 15:54:56 json_config -- scripts/common.sh@365 -- # ver1[v]=1 00:06:58.786 15:54:56 json_config -- scripts/common.sh@366 -- # decimal 2 00:06:58.786 15:54:56 json_config -- scripts/common.sh@353 -- # local d=2 00:06:58.787 15:54:56 json_config -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:58.787 15:54:56 json_config -- scripts/common.sh@355 -- # echo 2 00:06:58.787 15:54:56 json_config -- scripts/common.sh@366 -- # ver2[v]=2 00:06:58.787 15:54:56 json_config -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:58.787 15:54:56 json_config -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:58.787 15:54:56 json_config -- scripts/common.sh@368 -- # return 0 00:06:58.787 15:54:56 json_config -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:58.787 15:54:56 json_config -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:58.787 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.787 --rc genhtml_branch_coverage=1 00:06:58.787 --rc genhtml_function_coverage=1 00:06:58.787 --rc genhtml_legend=1 00:06:58.787 --rc geninfo_all_blocks=1 00:06:58.787 --rc geninfo_unexecuted_blocks=1 00:06:58.787 00:06:58.787 ' 00:06:58.787 15:54:56 json_config -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:58.787 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.787 --rc genhtml_branch_coverage=1 00:06:58.787 --rc genhtml_function_coverage=1 00:06:58.787 --rc genhtml_legend=1 00:06:58.787 --rc geninfo_all_blocks=1 00:06:58.787 --rc geninfo_unexecuted_blocks=1 00:06:58.787 00:06:58.787 ' 00:06:58.787 15:54:56 json_config -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:58.787 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.787 --rc genhtml_branch_coverage=1 00:06:58.787 --rc genhtml_function_coverage=1 00:06:58.787 --rc genhtml_legend=1 00:06:58.787 --rc geninfo_all_blocks=1 00:06:58.787 --rc geninfo_unexecuted_blocks=1 00:06:58.787 00:06:58.787 ' 00:06:58.787 15:54:56 json_config -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:58.787 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.787 --rc genhtml_branch_coverage=1 00:06:58.787 --rc genhtml_function_coverage=1 00:06:58.787 --rc genhtml_legend=1 00:06:58.787 --rc geninfo_all_blocks=1 00:06:58.787 --rc geninfo_unexecuted_blocks=1 00:06:58.787 00:06:58.787 ' 00:06:58.787 15:54:56 json_config -- json_config/json_config.sh@8 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@7 -- # uname -s 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:13ef8e52-d220-4442-baed-0e105827d1f9 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=13ef8e52-d220-4442-baed-0e105827d1f9 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:06:58.787 15:54:56 json_config -- scripts/common.sh@15 -- # shopt -s extglob 00:06:58.787 15:54:56 json_config -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:58.787 15:54:56 json_config -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:58.787 15:54:56 json_config -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:58.787 15:54:56 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:58.787 15:54:56 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:58.787 15:54:56 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:58.787 15:54:56 json_config -- paths/export.sh@5 -- # export PATH 00:06:58.787 15:54:56 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@51 -- # : 0 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:58.787 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:58.787 15:54:56 json_config -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:58.787 15:54:57 json_config -- json_config/json_config.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:06:58.787 15:54:57 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:06:58.787 15:54:57 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:06:58.787 15:54:57 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:06:58.787 15:54:57 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:06:58.787 WARNING: No tests are enabled so not running JSON configuration tests 00:06:58.787 15:54:57 json_config -- json_config/json_config.sh@27 -- # echo 'WARNING: No tests are enabled so not running JSON configuration tests' 00:06:58.787 15:54:57 json_config -- json_config/json_config.sh@28 -- # exit 0 00:06:58.787 00:06:58.787 real 0m0.132s 00:06:58.787 user 0m0.088s 00:06:58.787 sys 0m0.046s 00:06:58.787 15:54:57 json_config -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:58.787 15:54:57 json_config -- common/autotest_common.sh@10 -- # set +x 00:06:58.787 ************************************ 00:06:58.787 END TEST json_config 00:06:58.787 ************************************ 00:06:58.787 15:54:57 -- spdk/autotest.sh@160 -- # run_test json_config_extra_key /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:06:58.787 15:54:57 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:58.787 15:54:57 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:58.787 15:54:57 -- common/autotest_common.sh@10 -- # set +x 00:06:59.045 ************************************ 00:06:59.045 START TEST json_config_extra_key 00:06:59.045 ************************************ 00:06:59.045 15:54:57 json_config_extra_key -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:06:59.045 15:54:57 json_config_extra_key -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:59.045 15:54:57 json_config_extra_key -- common/autotest_common.sh@1693 -- # lcov --version 00:06:59.045 15:54:57 json_config_extra_key -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:59.045 15:54:57 json_config_extra_key -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@336 -- # IFS=.-: 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@336 -- # read -ra ver1 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@337 -- # IFS=.-: 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@337 -- # read -ra ver2 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@338 -- # local 'op=<' 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@340 -- # ver1_l=2 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@341 -- # ver2_l=1 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@344 -- # case "$op" in 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@345 -- # : 1 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@365 -- # decimal 1 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@353 -- # local d=1 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@355 -- # echo 1 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@365 -- # ver1[v]=1 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@366 -- # decimal 2 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@353 -- # local d=2 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@355 -- # echo 2 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@366 -- # ver2[v]=2 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:59.045 15:54:57 json_config_extra_key -- scripts/common.sh@368 -- # return 0 00:06:59.045 15:54:57 json_config_extra_key -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:59.045 15:54:57 json_config_extra_key -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:59.045 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:59.045 --rc genhtml_branch_coverage=1 00:06:59.045 --rc genhtml_function_coverage=1 00:06:59.045 --rc genhtml_legend=1 00:06:59.045 --rc geninfo_all_blocks=1 00:06:59.045 --rc geninfo_unexecuted_blocks=1 00:06:59.045 00:06:59.045 ' 00:06:59.045 15:54:57 json_config_extra_key -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:59.045 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:59.045 --rc genhtml_branch_coverage=1 00:06:59.045 --rc genhtml_function_coverage=1 00:06:59.045 --rc genhtml_legend=1 00:06:59.045 --rc geninfo_all_blocks=1 00:06:59.045 --rc geninfo_unexecuted_blocks=1 00:06:59.045 00:06:59.045 ' 00:06:59.045 15:54:57 json_config_extra_key -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:59.045 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:59.045 --rc genhtml_branch_coverage=1 00:06:59.045 --rc genhtml_function_coverage=1 00:06:59.045 --rc genhtml_legend=1 00:06:59.045 --rc geninfo_all_blocks=1 00:06:59.045 --rc geninfo_unexecuted_blocks=1 00:06:59.045 00:06:59.045 ' 00:06:59.045 15:54:57 json_config_extra_key -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:59.045 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:59.045 --rc genhtml_branch_coverage=1 00:06:59.045 --rc genhtml_function_coverage=1 00:06:59.045 --rc genhtml_legend=1 00:06:59.045 --rc geninfo_all_blocks=1 00:06:59.045 --rc geninfo_unexecuted_blocks=1 00:06:59.045 00:06:59.045 ' 00:06:59.045 15:54:57 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:06:59.045 15:54:57 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:06:59.045 15:54:57 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:59.045 15:54:57 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:59.045 15:54:57 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:59.045 15:54:57 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:59.045 15:54:57 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:59.045 15:54:57 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:59.045 15:54:57 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:59.045 15:54:57 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:59.045 15:54:57 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:59.045 15:54:57 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:59.045 15:54:57 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:13ef8e52-d220-4442-baed-0e105827d1f9 00:06:59.045 15:54:57 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=13ef8e52-d220-4442-baed-0e105827d1f9 00:06:59.045 15:54:57 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:59.046 15:54:57 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:59.046 15:54:57 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:06:59.046 15:54:57 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:59.046 15:54:57 json_config_extra_key -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:06:59.046 15:54:57 json_config_extra_key -- scripts/common.sh@15 -- # shopt -s extglob 00:06:59.046 15:54:57 json_config_extra_key -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:59.046 15:54:57 json_config_extra_key -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:59.046 15:54:57 json_config_extra_key -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:59.046 15:54:57 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:59.046 15:54:57 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:59.046 15:54:57 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:59.046 15:54:57 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:06:59.046 15:54:57 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:59.046 15:54:57 json_config_extra_key -- nvmf/common.sh@51 -- # : 0 00:06:59.046 15:54:57 json_config_extra_key -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:59.046 15:54:57 json_config_extra_key -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:59.046 15:54:57 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:59.046 15:54:57 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:59.046 15:54:57 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:59.046 15:54:57 json_config_extra_key -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:59.046 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:59.046 15:54:57 json_config_extra_key -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:59.046 15:54:57 json_config_extra_key -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:59.046 15:54:57 json_config_extra_key -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:59.046 15:54:57 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:06:59.046 15:54:57 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:06:59.046 15:54:57 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:06:59.046 15:54:57 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:06:59.046 15:54:57 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:06:59.046 15:54:57 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:06:59.046 15:54:57 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:06:59.046 15:54:57 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json') 00:06:59.046 15:54:57 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:06:59.046 INFO: launching applications... 00:06:59.046 15:54:57 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:06:59.046 15:54:57 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:06:59.046 15:54:57 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:06:59.046 15:54:57 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:06:59.046 15:54:57 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:06:59.046 15:54:57 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:06:59.046 15:54:57 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:06:59.046 15:54:57 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:06:59.046 15:54:57 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:06:59.046 15:54:57 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:06:59.046 15:54:57 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=56765 00:06:59.046 Waiting for target to run... 00:06:59.046 15:54:57 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:06:59.046 15:54:57 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 56765 /var/tmp/spdk_tgt.sock 00:06:59.046 15:54:57 json_config_extra_key -- common/autotest_common.sh@835 -- # '[' -z 56765 ']' 00:06:59.046 15:54:57 json_config_extra_key -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:06:59.046 15:54:57 json_config_extra_key -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:59.046 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:06:59.046 15:54:57 json_config_extra_key -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:06:59.046 15:54:57 json_config_extra_key -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:59.046 15:54:57 json_config_extra_key -- json_config/common.sh@21 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:06:59.046 15:54:57 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:06:59.046 [2024-11-20 15:54:57.249074] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:06:59.046 [2024-11-20 15:54:57.249197] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid56765 ] 00:06:59.611 [2024-11-20 15:54:57.580142] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:59.611 [2024-11-20 15:54:57.671556] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:00.175 15:54:58 json_config_extra_key -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:00.175 00:07:00.175 15:54:58 json_config_extra_key -- common/autotest_common.sh@868 -- # return 0 00:07:00.175 15:54:58 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:07:00.175 INFO: shutting down applications... 00:07:00.175 15:54:58 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:07:00.175 15:54:58 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:07:00.175 15:54:58 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:07:00.175 15:54:58 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:07:00.175 15:54:58 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 56765 ]] 00:07:00.175 15:54:58 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 56765 00:07:00.175 15:54:58 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:07:00.175 15:54:58 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:07:00.175 15:54:58 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 56765 00:07:00.175 15:54:58 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:07:00.433 15:54:58 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:07:00.433 15:54:58 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:07:00.433 15:54:58 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 56765 00:07:00.433 15:54:58 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:07:00.997 15:54:59 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:07:00.997 15:54:59 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:07:00.997 15:54:59 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 56765 00:07:00.997 15:54:59 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:07:01.560 15:54:59 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:07:01.560 15:54:59 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:07:01.560 15:54:59 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 56765 00:07:01.560 15:54:59 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:07:02.123 15:55:00 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:07:02.123 15:55:00 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:07:02.123 15:55:00 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 56765 00:07:02.123 15:55:00 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:07:02.123 15:55:00 json_config_extra_key -- json_config/common.sh@43 -- # break 00:07:02.123 15:55:00 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:07:02.123 SPDK target shutdown done 00:07:02.123 15:55:00 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:07:02.123 Success 00:07:02.123 15:55:00 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:07:02.123 00:07:02.123 real 0m3.157s 00:07:02.123 user 0m2.761s 00:07:02.123 sys 0m0.401s 00:07:02.123 15:55:00 json_config_extra_key -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:02.123 15:55:00 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:07:02.123 ************************************ 00:07:02.123 END TEST json_config_extra_key 00:07:02.123 ************************************ 00:07:02.123 15:55:00 -- spdk/autotest.sh@161 -- # run_test alias_rpc /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:07:02.123 15:55:00 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:02.123 15:55:00 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:02.123 15:55:00 -- common/autotest_common.sh@10 -- # set +x 00:07:02.123 ************************************ 00:07:02.123 START TEST alias_rpc 00:07:02.123 ************************************ 00:07:02.123 15:55:00 alias_rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:07:02.123 * Looking for test storage... 00:07:02.123 * Found test storage at /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc 00:07:02.123 15:55:00 alias_rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:07:02.123 15:55:00 alias_rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:07:02.123 15:55:00 alias_rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:07:02.123 15:55:00 alias_rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@344 -- # case "$op" in 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@345 -- # : 1 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@365 -- # decimal 1 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@353 -- # local d=1 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@355 -- # echo 1 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@366 -- # decimal 2 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@353 -- # local d=2 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:02.123 15:55:00 alias_rpc -- scripts/common.sh@355 -- # echo 2 00:07:02.380 15:55:00 alias_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:07:02.380 15:55:00 alias_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:02.380 15:55:00 alias_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:02.380 15:55:00 alias_rpc -- scripts/common.sh@368 -- # return 0 00:07:02.380 15:55:00 alias_rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:02.380 15:55:00 alias_rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:07:02.380 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:02.380 --rc genhtml_branch_coverage=1 00:07:02.380 --rc genhtml_function_coverage=1 00:07:02.380 --rc genhtml_legend=1 00:07:02.380 --rc geninfo_all_blocks=1 00:07:02.380 --rc geninfo_unexecuted_blocks=1 00:07:02.380 00:07:02.380 ' 00:07:02.380 15:55:00 alias_rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:07:02.380 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:02.380 --rc genhtml_branch_coverage=1 00:07:02.380 --rc genhtml_function_coverage=1 00:07:02.380 --rc genhtml_legend=1 00:07:02.380 --rc geninfo_all_blocks=1 00:07:02.380 --rc geninfo_unexecuted_blocks=1 00:07:02.380 00:07:02.380 ' 00:07:02.380 15:55:00 alias_rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:07:02.380 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:02.380 --rc genhtml_branch_coverage=1 00:07:02.380 --rc genhtml_function_coverage=1 00:07:02.380 --rc genhtml_legend=1 00:07:02.380 --rc geninfo_all_blocks=1 00:07:02.380 --rc geninfo_unexecuted_blocks=1 00:07:02.380 00:07:02.380 ' 00:07:02.380 15:55:00 alias_rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:07:02.380 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:02.380 --rc genhtml_branch_coverage=1 00:07:02.380 --rc genhtml_function_coverage=1 00:07:02.380 --rc genhtml_legend=1 00:07:02.380 --rc geninfo_all_blocks=1 00:07:02.380 --rc geninfo_unexecuted_blocks=1 00:07:02.380 00:07:02.380 ' 00:07:02.380 15:55:00 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:07:02.380 15:55:00 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=56858 00:07:02.380 15:55:00 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 56858 00:07:02.380 15:55:00 alias_rpc -- common/autotest_common.sh@835 -- # '[' -z 56858 ']' 00:07:02.380 15:55:00 alias_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:02.380 15:55:00 alias_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:02.380 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:02.380 15:55:00 alias_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:02.380 15:55:00 alias_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:02.380 15:55:00 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:07:02.380 15:55:00 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:02.380 [2024-11-20 15:55:00.446632] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:07:02.380 [2024-11-20 15:55:00.446769] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid56858 ] 00:07:02.380 [2024-11-20 15:55:00.597723] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:02.639 [2024-11-20 15:55:00.694375] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:03.203 15:55:01 alias_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:03.203 15:55:01 alias_rpc -- common/autotest_common.sh@868 -- # return 0 00:07:03.203 15:55:01 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py load_config -i 00:07:03.460 15:55:01 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 56858 00:07:03.460 15:55:01 alias_rpc -- common/autotest_common.sh@954 -- # '[' -z 56858 ']' 00:07:03.460 15:55:01 alias_rpc -- common/autotest_common.sh@958 -- # kill -0 56858 00:07:03.460 15:55:01 alias_rpc -- common/autotest_common.sh@959 -- # uname 00:07:03.460 15:55:01 alias_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:03.460 15:55:01 alias_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 56858 00:07:03.460 15:55:01 alias_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:03.460 15:55:01 alias_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:03.460 killing process with pid 56858 00:07:03.460 15:55:01 alias_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 56858' 00:07:03.460 15:55:01 alias_rpc -- common/autotest_common.sh@973 -- # kill 56858 00:07:03.460 15:55:01 alias_rpc -- common/autotest_common.sh@978 -- # wait 56858 00:07:04.834 00:07:04.834 real 0m2.847s 00:07:04.834 user 0m2.961s 00:07:04.834 sys 0m0.395s 00:07:04.834 15:55:03 alias_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:04.834 15:55:03 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:04.834 ************************************ 00:07:04.834 END TEST alias_rpc 00:07:04.834 ************************************ 00:07:05.092 15:55:03 -- spdk/autotest.sh@163 -- # [[ 0 -eq 0 ]] 00:07:05.092 15:55:03 -- spdk/autotest.sh@164 -- # run_test spdkcli_tcp /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:07:05.092 15:55:03 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:05.092 15:55:03 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:05.092 15:55:03 -- common/autotest_common.sh@10 -- # set +x 00:07:05.092 ************************************ 00:07:05.092 START TEST spdkcli_tcp 00:07:05.092 ************************************ 00:07:05.092 15:55:03 spdkcli_tcp -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:07:05.092 * Looking for test storage... 00:07:05.092 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:07:05.092 15:55:03 spdkcli_tcp -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:07:05.092 15:55:03 spdkcli_tcp -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:07:05.092 15:55:03 spdkcli_tcp -- common/autotest_common.sh@1693 -- # lcov --version 00:07:05.092 15:55:03 spdkcli_tcp -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@344 -- # case "$op" in 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@345 -- # : 1 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@365 -- # decimal 1 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@353 -- # local d=1 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@355 -- # echo 1 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@366 -- # decimal 2 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@353 -- # local d=2 00:07:05.092 15:55:03 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:05.093 15:55:03 spdkcli_tcp -- scripts/common.sh@355 -- # echo 2 00:07:05.093 15:55:03 spdkcli_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:07:05.093 15:55:03 spdkcli_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:05.093 15:55:03 spdkcli_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:05.093 15:55:03 spdkcli_tcp -- scripts/common.sh@368 -- # return 0 00:07:05.093 15:55:03 spdkcli_tcp -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:05.093 15:55:03 spdkcli_tcp -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:07:05.093 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:05.093 --rc genhtml_branch_coverage=1 00:07:05.093 --rc genhtml_function_coverage=1 00:07:05.093 --rc genhtml_legend=1 00:07:05.093 --rc geninfo_all_blocks=1 00:07:05.093 --rc geninfo_unexecuted_blocks=1 00:07:05.093 00:07:05.093 ' 00:07:05.093 15:55:03 spdkcli_tcp -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:07:05.093 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:05.093 --rc genhtml_branch_coverage=1 00:07:05.093 --rc genhtml_function_coverage=1 00:07:05.093 --rc genhtml_legend=1 00:07:05.093 --rc geninfo_all_blocks=1 00:07:05.093 --rc geninfo_unexecuted_blocks=1 00:07:05.093 00:07:05.093 ' 00:07:05.093 15:55:03 spdkcli_tcp -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:07:05.093 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:05.093 --rc genhtml_branch_coverage=1 00:07:05.093 --rc genhtml_function_coverage=1 00:07:05.093 --rc genhtml_legend=1 00:07:05.093 --rc geninfo_all_blocks=1 00:07:05.093 --rc geninfo_unexecuted_blocks=1 00:07:05.093 00:07:05.093 ' 00:07:05.093 15:55:03 spdkcli_tcp -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:07:05.093 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:05.093 --rc genhtml_branch_coverage=1 00:07:05.093 --rc genhtml_function_coverage=1 00:07:05.093 --rc genhtml_legend=1 00:07:05.093 --rc geninfo_all_blocks=1 00:07:05.093 --rc geninfo_unexecuted_blocks=1 00:07:05.093 00:07:05.093 ' 00:07:05.093 15:55:03 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:07:05.093 15:55:03 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:07:05.093 15:55:03 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:07:05.093 15:55:03 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:07:05.093 15:55:03 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:07:05.093 15:55:03 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:07:05.093 15:55:03 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:07:05.093 15:55:03 spdkcli_tcp -- common/autotest_common.sh@726 -- # xtrace_disable 00:07:05.093 15:55:03 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:07:05.093 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:05.093 15:55:03 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=56954 00:07:05.093 15:55:03 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 56954 00:07:05.093 15:55:03 spdkcli_tcp -- common/autotest_common.sh@835 -- # '[' -z 56954 ']' 00:07:05.093 15:55:03 spdkcli_tcp -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:05.093 15:55:03 spdkcli_tcp -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:05.093 15:55:03 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:07:05.093 15:55:03 spdkcli_tcp -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:05.093 15:55:03 spdkcli_tcp -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:05.093 15:55:03 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:07:05.093 [2024-11-20 15:55:03.337325] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:07:05.093 [2024-11-20 15:55:03.337450] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid56954 ] 00:07:05.350 [2024-11-20 15:55:03.497740] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:07:05.608 [2024-11-20 15:55:03.613994] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:05.608 [2024-11-20 15:55:03.614006] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:06.174 15:55:04 spdkcli_tcp -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:06.174 15:55:04 spdkcli_tcp -- common/autotest_common.sh@868 -- # return 0 00:07:06.174 15:55:04 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=56971 00:07:06.174 15:55:04 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:07:06.174 15:55:04 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:07:06.432 [ 00:07:06.432 "bdev_malloc_delete", 00:07:06.432 "bdev_malloc_create", 00:07:06.432 "bdev_null_resize", 00:07:06.432 "bdev_null_delete", 00:07:06.432 "bdev_null_create", 00:07:06.432 "bdev_nvme_cuse_unregister", 00:07:06.432 "bdev_nvme_cuse_register", 00:07:06.432 "bdev_opal_new_user", 00:07:06.432 "bdev_opal_set_lock_state", 00:07:06.432 "bdev_opal_delete", 00:07:06.432 "bdev_opal_get_info", 00:07:06.433 "bdev_opal_create", 00:07:06.433 "bdev_nvme_opal_revert", 00:07:06.433 "bdev_nvme_opal_init", 00:07:06.433 "bdev_nvme_send_cmd", 00:07:06.433 "bdev_nvme_set_keys", 00:07:06.433 "bdev_nvme_get_path_iostat", 00:07:06.433 "bdev_nvme_get_mdns_discovery_info", 00:07:06.433 "bdev_nvme_stop_mdns_discovery", 00:07:06.433 "bdev_nvme_start_mdns_discovery", 00:07:06.433 "bdev_nvme_set_multipath_policy", 00:07:06.433 "bdev_nvme_set_preferred_path", 00:07:06.433 "bdev_nvme_get_io_paths", 00:07:06.433 "bdev_nvme_remove_error_injection", 00:07:06.433 "bdev_nvme_add_error_injection", 00:07:06.433 "bdev_nvme_get_discovery_info", 00:07:06.433 "bdev_nvme_stop_discovery", 00:07:06.433 "bdev_nvme_start_discovery", 00:07:06.433 "bdev_nvme_get_controller_health_info", 00:07:06.433 "bdev_nvme_disable_controller", 00:07:06.433 "bdev_nvme_enable_controller", 00:07:06.433 "bdev_nvme_reset_controller", 00:07:06.433 "bdev_nvme_get_transport_statistics", 00:07:06.433 "bdev_nvme_apply_firmware", 00:07:06.433 "bdev_nvme_detach_controller", 00:07:06.433 "bdev_nvme_get_controllers", 00:07:06.433 "bdev_nvme_attach_controller", 00:07:06.433 "bdev_nvme_set_hotplug", 00:07:06.433 "bdev_nvme_set_options", 00:07:06.433 "bdev_passthru_delete", 00:07:06.433 "bdev_passthru_create", 00:07:06.433 "bdev_lvol_set_parent_bdev", 00:07:06.433 "bdev_lvol_set_parent", 00:07:06.433 "bdev_lvol_check_shallow_copy", 00:07:06.433 "bdev_lvol_start_shallow_copy", 00:07:06.433 "bdev_lvol_grow_lvstore", 00:07:06.433 "bdev_lvol_get_lvols", 00:07:06.433 "bdev_lvol_get_lvstores", 00:07:06.433 "bdev_lvol_delete", 00:07:06.433 "bdev_lvol_set_read_only", 00:07:06.433 "bdev_lvol_resize", 00:07:06.433 "bdev_lvol_decouple_parent", 00:07:06.433 "bdev_lvol_inflate", 00:07:06.433 "bdev_lvol_rename", 00:07:06.433 "bdev_lvol_clone_bdev", 00:07:06.433 "bdev_lvol_clone", 00:07:06.433 "bdev_lvol_snapshot", 00:07:06.433 "bdev_lvol_create", 00:07:06.433 "bdev_lvol_delete_lvstore", 00:07:06.433 "bdev_lvol_rename_lvstore", 00:07:06.433 "bdev_lvol_create_lvstore", 00:07:06.433 "bdev_raid_set_options", 00:07:06.433 "bdev_raid_remove_base_bdev", 00:07:06.433 "bdev_raid_add_base_bdev", 00:07:06.433 "bdev_raid_delete", 00:07:06.433 "bdev_raid_create", 00:07:06.433 "bdev_raid_get_bdevs", 00:07:06.433 "bdev_error_inject_error", 00:07:06.433 "bdev_error_delete", 00:07:06.433 "bdev_error_create", 00:07:06.433 "bdev_split_delete", 00:07:06.433 "bdev_split_create", 00:07:06.433 "bdev_delay_delete", 00:07:06.433 "bdev_delay_create", 00:07:06.433 "bdev_delay_update_latency", 00:07:06.433 "bdev_zone_block_delete", 00:07:06.433 "bdev_zone_block_create", 00:07:06.433 "blobfs_create", 00:07:06.433 "blobfs_detect", 00:07:06.433 "blobfs_set_cache_size", 00:07:06.433 "bdev_aio_delete", 00:07:06.433 "bdev_aio_rescan", 00:07:06.433 "bdev_aio_create", 00:07:06.433 "bdev_ftl_set_property", 00:07:06.433 "bdev_ftl_get_properties", 00:07:06.433 "bdev_ftl_get_stats", 00:07:06.433 "bdev_ftl_unmap", 00:07:06.433 "bdev_ftl_unload", 00:07:06.433 "bdev_ftl_delete", 00:07:06.433 "bdev_ftl_load", 00:07:06.433 "bdev_ftl_create", 00:07:06.433 "bdev_virtio_attach_controller", 00:07:06.433 "bdev_virtio_scsi_get_devices", 00:07:06.433 "bdev_virtio_detach_controller", 00:07:06.433 "bdev_virtio_blk_set_hotplug", 00:07:06.433 "bdev_iscsi_delete", 00:07:06.433 "bdev_iscsi_create", 00:07:06.433 "bdev_iscsi_set_options", 00:07:06.433 "accel_error_inject_error", 00:07:06.433 "ioat_scan_accel_module", 00:07:06.433 "dsa_scan_accel_module", 00:07:06.433 "iaa_scan_accel_module", 00:07:06.433 "keyring_file_remove_key", 00:07:06.433 "keyring_file_add_key", 00:07:06.433 "keyring_linux_set_options", 00:07:06.433 "fsdev_aio_delete", 00:07:06.433 "fsdev_aio_create", 00:07:06.433 "iscsi_get_histogram", 00:07:06.433 "iscsi_enable_histogram", 00:07:06.433 "iscsi_set_options", 00:07:06.433 "iscsi_get_auth_groups", 00:07:06.433 "iscsi_auth_group_remove_secret", 00:07:06.433 "iscsi_auth_group_add_secret", 00:07:06.433 "iscsi_delete_auth_group", 00:07:06.433 "iscsi_create_auth_group", 00:07:06.433 "iscsi_set_discovery_auth", 00:07:06.433 "iscsi_get_options", 00:07:06.433 "iscsi_target_node_request_logout", 00:07:06.433 "iscsi_target_node_set_redirect", 00:07:06.433 "iscsi_target_node_set_auth", 00:07:06.433 "iscsi_target_node_add_lun", 00:07:06.433 "iscsi_get_stats", 00:07:06.433 "iscsi_get_connections", 00:07:06.433 "iscsi_portal_group_set_auth", 00:07:06.433 "iscsi_start_portal_group", 00:07:06.433 "iscsi_delete_portal_group", 00:07:06.433 "iscsi_create_portal_group", 00:07:06.433 "iscsi_get_portal_groups", 00:07:06.433 "iscsi_delete_target_node", 00:07:06.433 "iscsi_target_node_remove_pg_ig_maps", 00:07:06.433 "iscsi_target_node_add_pg_ig_maps", 00:07:06.433 "iscsi_create_target_node", 00:07:06.433 "iscsi_get_target_nodes", 00:07:06.433 "iscsi_delete_initiator_group", 00:07:06.433 "iscsi_initiator_group_remove_initiators", 00:07:06.433 "iscsi_initiator_group_add_initiators", 00:07:06.433 "iscsi_create_initiator_group", 00:07:06.433 "iscsi_get_initiator_groups", 00:07:06.433 "nvmf_set_crdt", 00:07:06.433 "nvmf_set_config", 00:07:06.433 "nvmf_set_max_subsystems", 00:07:06.433 "nvmf_stop_mdns_prr", 00:07:06.433 "nvmf_publish_mdns_prr", 00:07:06.433 "nvmf_subsystem_get_listeners", 00:07:06.433 "nvmf_subsystem_get_qpairs", 00:07:06.433 "nvmf_subsystem_get_controllers", 00:07:06.433 "nvmf_get_stats", 00:07:06.433 "nvmf_get_transports", 00:07:06.433 "nvmf_create_transport", 00:07:06.433 "nvmf_get_targets", 00:07:06.433 "nvmf_delete_target", 00:07:06.433 "nvmf_create_target", 00:07:06.433 "nvmf_subsystem_allow_any_host", 00:07:06.433 "nvmf_subsystem_set_keys", 00:07:06.433 "nvmf_subsystem_remove_host", 00:07:06.433 "nvmf_subsystem_add_host", 00:07:06.433 "nvmf_ns_remove_host", 00:07:06.433 "nvmf_ns_add_host", 00:07:06.433 "nvmf_subsystem_remove_ns", 00:07:06.433 "nvmf_subsystem_set_ns_ana_group", 00:07:06.433 "nvmf_subsystem_add_ns", 00:07:06.433 "nvmf_subsystem_listener_set_ana_state", 00:07:06.433 "nvmf_discovery_get_referrals", 00:07:06.433 "nvmf_discovery_remove_referral", 00:07:06.433 "nvmf_discovery_add_referral", 00:07:06.433 "nvmf_subsystem_remove_listener", 00:07:06.433 "nvmf_subsystem_add_listener", 00:07:06.433 "nvmf_delete_subsystem", 00:07:06.433 "nvmf_create_subsystem", 00:07:06.433 "nvmf_get_subsystems", 00:07:06.433 "env_dpdk_get_mem_stats", 00:07:06.433 "nbd_get_disks", 00:07:06.433 "nbd_stop_disk", 00:07:06.433 "nbd_start_disk", 00:07:06.433 "ublk_recover_disk", 00:07:06.433 "ublk_get_disks", 00:07:06.433 "ublk_stop_disk", 00:07:06.433 "ublk_start_disk", 00:07:06.433 "ublk_destroy_target", 00:07:06.433 "ublk_create_target", 00:07:06.433 "virtio_blk_create_transport", 00:07:06.433 "virtio_blk_get_transports", 00:07:06.433 "vhost_controller_set_coalescing", 00:07:06.433 "vhost_get_controllers", 00:07:06.433 "vhost_delete_controller", 00:07:06.433 "vhost_create_blk_controller", 00:07:06.433 "vhost_scsi_controller_remove_target", 00:07:06.433 "vhost_scsi_controller_add_target", 00:07:06.433 "vhost_start_scsi_controller", 00:07:06.433 "vhost_create_scsi_controller", 00:07:06.433 "thread_set_cpumask", 00:07:06.433 "scheduler_set_options", 00:07:06.433 "framework_get_governor", 00:07:06.433 "framework_get_scheduler", 00:07:06.433 "framework_set_scheduler", 00:07:06.433 "framework_get_reactors", 00:07:06.433 "thread_get_io_channels", 00:07:06.433 "thread_get_pollers", 00:07:06.433 "thread_get_stats", 00:07:06.433 "framework_monitor_context_switch", 00:07:06.433 "spdk_kill_instance", 00:07:06.433 "log_enable_timestamps", 00:07:06.433 "log_get_flags", 00:07:06.433 "log_clear_flag", 00:07:06.433 "log_set_flag", 00:07:06.433 "log_get_level", 00:07:06.433 "log_set_level", 00:07:06.433 "log_get_print_level", 00:07:06.433 "log_set_print_level", 00:07:06.433 "framework_enable_cpumask_locks", 00:07:06.433 "framework_disable_cpumask_locks", 00:07:06.433 "framework_wait_init", 00:07:06.433 "framework_start_init", 00:07:06.433 "scsi_get_devices", 00:07:06.433 "bdev_get_histogram", 00:07:06.433 "bdev_enable_histogram", 00:07:06.433 "bdev_set_qos_limit", 00:07:06.433 "bdev_set_qd_sampling_period", 00:07:06.433 "bdev_get_bdevs", 00:07:06.433 "bdev_reset_iostat", 00:07:06.433 "bdev_get_iostat", 00:07:06.433 "bdev_examine", 00:07:06.433 "bdev_wait_for_examine", 00:07:06.433 "bdev_set_options", 00:07:06.433 "accel_get_stats", 00:07:06.433 "accel_set_options", 00:07:06.433 "accel_set_driver", 00:07:06.433 "accel_crypto_key_destroy", 00:07:06.433 "accel_crypto_keys_get", 00:07:06.433 "accel_crypto_key_create", 00:07:06.433 "accel_assign_opc", 00:07:06.433 "accel_get_module_info", 00:07:06.433 "accel_get_opc_assignments", 00:07:06.433 "vmd_rescan", 00:07:06.433 "vmd_remove_device", 00:07:06.433 "vmd_enable", 00:07:06.433 "sock_get_default_impl", 00:07:06.433 "sock_set_default_impl", 00:07:06.433 "sock_impl_set_options", 00:07:06.433 "sock_impl_get_options", 00:07:06.433 "iobuf_get_stats", 00:07:06.433 "iobuf_set_options", 00:07:06.433 "keyring_get_keys", 00:07:06.433 "framework_get_pci_devices", 00:07:06.433 "framework_get_config", 00:07:06.433 "framework_get_subsystems", 00:07:06.433 "fsdev_set_opts", 00:07:06.433 "fsdev_get_opts", 00:07:06.433 "trace_get_info", 00:07:06.433 "trace_get_tpoint_group_mask", 00:07:06.434 "trace_disable_tpoint_group", 00:07:06.434 "trace_enable_tpoint_group", 00:07:06.434 "trace_clear_tpoint_mask", 00:07:06.434 "trace_set_tpoint_mask", 00:07:06.434 "notify_get_notifications", 00:07:06.434 "notify_get_types", 00:07:06.434 "spdk_get_version", 00:07:06.434 "rpc_get_methods" 00:07:06.434 ] 00:07:06.434 15:55:04 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:07:06.434 15:55:04 spdkcli_tcp -- common/autotest_common.sh@732 -- # xtrace_disable 00:07:06.434 15:55:04 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:07:06.434 15:55:04 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:07:06.434 15:55:04 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 56954 00:07:06.434 15:55:04 spdkcli_tcp -- common/autotest_common.sh@954 -- # '[' -z 56954 ']' 00:07:06.434 15:55:04 spdkcli_tcp -- common/autotest_common.sh@958 -- # kill -0 56954 00:07:06.434 15:55:04 spdkcli_tcp -- common/autotest_common.sh@959 -- # uname 00:07:06.434 15:55:04 spdkcli_tcp -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:06.434 15:55:04 spdkcli_tcp -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 56954 00:07:06.434 15:55:04 spdkcli_tcp -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:06.434 15:55:04 spdkcli_tcp -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:06.434 15:55:04 spdkcli_tcp -- common/autotest_common.sh@972 -- # echo 'killing process with pid 56954' 00:07:06.434 killing process with pid 56954 00:07:06.434 15:55:04 spdkcli_tcp -- common/autotest_common.sh@973 -- # kill 56954 00:07:06.434 15:55:04 spdkcli_tcp -- common/autotest_common.sh@978 -- # wait 56954 00:07:08.334 00:07:08.334 real 0m2.982s 00:07:08.334 user 0m5.423s 00:07:08.334 sys 0m0.449s 00:07:08.334 15:55:06 spdkcli_tcp -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:08.334 ************************************ 00:07:08.334 END TEST spdkcli_tcp 00:07:08.334 ************************************ 00:07:08.334 15:55:06 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:07:08.334 15:55:06 -- spdk/autotest.sh@167 -- # run_test dpdk_mem_utility /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:07:08.334 15:55:06 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:08.334 15:55:06 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:08.334 15:55:06 -- common/autotest_common.sh@10 -- # set +x 00:07:08.334 ************************************ 00:07:08.334 START TEST dpdk_mem_utility 00:07:08.334 ************************************ 00:07:08.334 15:55:06 dpdk_mem_utility -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:07:08.334 * Looking for test storage... 00:07:08.334 * Found test storage at /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility 00:07:08.334 15:55:06 dpdk_mem_utility -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:07:08.334 15:55:06 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # lcov --version 00:07:08.334 15:55:06 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:07:08.334 15:55:06 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@336 -- # IFS=.-: 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@336 -- # read -ra ver1 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@337 -- # IFS=.-: 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@337 -- # read -ra ver2 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@338 -- # local 'op=<' 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@340 -- # ver1_l=2 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@341 -- # ver2_l=1 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@344 -- # case "$op" in 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@345 -- # : 1 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@365 -- # decimal 1 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=1 00:07:08.334 15:55:06 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:08.335 15:55:06 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 1 00:07:08.335 15:55:06 dpdk_mem_utility -- scripts/common.sh@365 -- # ver1[v]=1 00:07:08.335 15:55:06 dpdk_mem_utility -- scripts/common.sh@366 -- # decimal 2 00:07:08.335 15:55:06 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=2 00:07:08.335 15:55:06 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:08.335 15:55:06 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 2 00:07:08.335 15:55:06 dpdk_mem_utility -- scripts/common.sh@366 -- # ver2[v]=2 00:07:08.335 15:55:06 dpdk_mem_utility -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:08.335 15:55:06 dpdk_mem_utility -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:08.335 15:55:06 dpdk_mem_utility -- scripts/common.sh@368 -- # return 0 00:07:08.335 15:55:06 dpdk_mem_utility -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:08.335 15:55:06 dpdk_mem_utility -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:07:08.335 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:08.335 --rc genhtml_branch_coverage=1 00:07:08.335 --rc genhtml_function_coverage=1 00:07:08.335 --rc genhtml_legend=1 00:07:08.335 --rc geninfo_all_blocks=1 00:07:08.335 --rc geninfo_unexecuted_blocks=1 00:07:08.335 00:07:08.335 ' 00:07:08.335 15:55:06 dpdk_mem_utility -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:07:08.335 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:08.335 --rc genhtml_branch_coverage=1 00:07:08.335 --rc genhtml_function_coverage=1 00:07:08.335 --rc genhtml_legend=1 00:07:08.335 --rc geninfo_all_blocks=1 00:07:08.335 --rc geninfo_unexecuted_blocks=1 00:07:08.335 00:07:08.335 ' 00:07:08.335 15:55:06 dpdk_mem_utility -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:07:08.335 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:08.335 --rc genhtml_branch_coverage=1 00:07:08.335 --rc genhtml_function_coverage=1 00:07:08.335 --rc genhtml_legend=1 00:07:08.335 --rc geninfo_all_blocks=1 00:07:08.335 --rc geninfo_unexecuted_blocks=1 00:07:08.335 00:07:08.335 ' 00:07:08.335 15:55:06 dpdk_mem_utility -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:07:08.335 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:08.335 --rc genhtml_branch_coverage=1 00:07:08.335 --rc genhtml_function_coverage=1 00:07:08.335 --rc genhtml_legend=1 00:07:08.335 --rc geninfo_all_blocks=1 00:07:08.335 --rc geninfo_unexecuted_blocks=1 00:07:08.335 00:07:08.335 ' 00:07:08.335 15:55:06 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:07:08.335 15:55:06 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:07:08.335 15:55:06 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=57065 00:07:08.335 15:55:06 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 57065 00:07:08.335 15:55:06 dpdk_mem_utility -- common/autotest_common.sh@835 -- # '[' -z 57065 ']' 00:07:08.335 15:55:06 dpdk_mem_utility -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:08.335 15:55:06 dpdk_mem_utility -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:08.335 15:55:06 dpdk_mem_utility -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:08.335 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:08.335 15:55:06 dpdk_mem_utility -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:08.335 15:55:06 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:07:08.335 [2024-11-20 15:55:06.349994] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:07:08.335 [2024-11-20 15:55:06.350103] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57065 ] 00:07:08.335 [2024-11-20 15:55:06.506768] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:08.592 [2024-11-20 15:55:06.609756] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:09.216 15:55:07 dpdk_mem_utility -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:09.216 15:55:07 dpdk_mem_utility -- common/autotest_common.sh@868 -- # return 0 00:07:09.216 15:55:07 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:07:09.216 15:55:07 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:07:09.216 15:55:07 dpdk_mem_utility -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:09.216 15:55:07 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:07:09.216 { 00:07:09.216 "filename": "/tmp/spdk_mem_dump.txt" 00:07:09.216 } 00:07:09.216 15:55:07 dpdk_mem_utility -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:09.216 15:55:07 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:07:09.216 DPDK memory size 824.000000 MiB in 1 heap(s) 00:07:09.216 1 heaps totaling size 824.000000 MiB 00:07:09.216 size: 824.000000 MiB heap id: 0 00:07:09.216 end heaps---------- 00:07:09.216 9 mempools totaling size 603.782043 MiB 00:07:09.216 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:07:09.216 size: 158.602051 MiB name: PDU_data_out_Pool 00:07:09.216 size: 100.555481 MiB name: bdev_io_57065 00:07:09.216 size: 50.003479 MiB name: msgpool_57065 00:07:09.216 size: 36.509338 MiB name: fsdev_io_57065 00:07:09.216 size: 21.763794 MiB name: PDU_Pool 00:07:09.216 size: 19.513306 MiB name: SCSI_TASK_Pool 00:07:09.216 size: 4.133484 MiB name: evtpool_57065 00:07:09.216 size: 0.026123 MiB name: Session_Pool 00:07:09.216 end mempools------- 00:07:09.216 6 memzones totaling size 4.142822 MiB 00:07:09.216 size: 1.000366 MiB name: RG_ring_0_57065 00:07:09.216 size: 1.000366 MiB name: RG_ring_1_57065 00:07:09.216 size: 1.000366 MiB name: RG_ring_4_57065 00:07:09.216 size: 1.000366 MiB name: RG_ring_5_57065 00:07:09.216 size: 0.125366 MiB name: RG_ring_2_57065 00:07:09.216 size: 0.015991 MiB name: RG_ring_3_57065 00:07:09.216 end memzones------- 00:07:09.216 15:55:07 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py -m 0 00:07:09.216 heap id: 0 total size: 824.000000 MiB number of busy elements: 307 number of free elements: 18 00:07:09.216 list of free elements. size: 16.783325 MiB 00:07:09.216 element at address: 0x200006400000 with size: 1.995972 MiB 00:07:09.216 element at address: 0x20000a600000 with size: 1.995972 MiB 00:07:09.216 element at address: 0x200003e00000 with size: 1.991028 MiB 00:07:09.216 element at address: 0x200019500040 with size: 0.999939 MiB 00:07:09.216 element at address: 0x200019900040 with size: 0.999939 MiB 00:07:09.216 element at address: 0x200019a00000 with size: 0.999084 MiB 00:07:09.216 element at address: 0x200032600000 with size: 0.994324 MiB 00:07:09.216 element at address: 0x200000400000 with size: 0.992004 MiB 00:07:09.216 element at address: 0x200019200000 with size: 0.959656 MiB 00:07:09.216 element at address: 0x200019d00040 with size: 0.936401 MiB 00:07:09.216 element at address: 0x200000200000 with size: 0.716980 MiB 00:07:09.216 element at address: 0x20001b400000 with size: 0.562683 MiB 00:07:09.216 element at address: 0x200000c00000 with size: 0.489929 MiB 00:07:09.216 element at address: 0x200019600000 with size: 0.487976 MiB 00:07:09.216 element at address: 0x200019e00000 with size: 0.485413 MiB 00:07:09.216 element at address: 0x200012c00000 with size: 0.433228 MiB 00:07:09.216 element at address: 0x200028800000 with size: 0.391907 MiB 00:07:09.216 element at address: 0x200000800000 with size: 0.350891 MiB 00:07:09.216 list of standard malloc elements. size: 199.285767 MiB 00:07:09.216 element at address: 0x20000a7fef80 with size: 132.000183 MiB 00:07:09.216 element at address: 0x2000065fef80 with size: 64.000183 MiB 00:07:09.216 element at address: 0x2000193fff80 with size: 1.000183 MiB 00:07:09.216 element at address: 0x2000197fff80 with size: 1.000183 MiB 00:07:09.216 element at address: 0x200019bfff80 with size: 1.000183 MiB 00:07:09.216 element at address: 0x2000003d9e80 with size: 0.140808 MiB 00:07:09.216 element at address: 0x200019deff40 with size: 0.062683 MiB 00:07:09.216 element at address: 0x2000003fdf40 with size: 0.007996 MiB 00:07:09.216 element at address: 0x20000a5ff040 with size: 0.000427 MiB 00:07:09.216 element at address: 0x200019defdc0 with size: 0.000366 MiB 00:07:09.216 element at address: 0x200012bff040 with size: 0.000305 MiB 00:07:09.216 element at address: 0x2000002d7b00 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000003d9d80 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004fdf40 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004fe040 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004fe140 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004fe240 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004fe340 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004fe440 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004fe540 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004fe640 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004fe740 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004fe840 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004fe940 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004fea40 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004feb40 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004fec40 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004fed40 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004fee40 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004fef40 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004ff040 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004ff140 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004ff240 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004ff340 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004ff440 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004ff540 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004ff640 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004ff740 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004ff840 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004ff940 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004ffbc0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004ffcc0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000004ffdc0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087e1c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087e2c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087e3c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087e4c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087e5c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087e6c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087e7c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087e8c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087e9c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087eac0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087ebc0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087ecc0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087edc0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087eec0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087efc0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087f0c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087f1c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087f2c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087f3c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x20000087f4c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000008ff800 with size: 0.000244 MiB 00:07:09.216 element at address: 0x2000008ffa80 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7d6c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7d7c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7d8c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7d9c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7dac0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7dbc0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7dcc0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7ddc0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7dec0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7dfc0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7e0c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7e1c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7e2c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7e3c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7e4c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7e5c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7e6c0 with size: 0.000244 MiB 00:07:09.216 element at address: 0x200000c7e7c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200000c7e8c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200000c7e9c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200000c7eac0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200000c7ebc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200000cfef00 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200000cff000 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20000a5ff200 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20000a5ff300 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20000a5ff400 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20000a5ff500 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20000a5ff600 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20000a5ff700 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20000a5ff800 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20000a5ff900 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20000a5ffa00 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20000a5ffb00 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20000a5ffc00 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20000a5ffd00 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20000a5ffe00 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20000a5fff00 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012bff180 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012bff280 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012bff380 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012bff480 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012bff580 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012bff680 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012bff780 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012bff880 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012bff980 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012bffa80 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012bffb80 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012bffc80 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012bfff00 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012c6ee80 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012c6ef80 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012c6f080 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012c6f180 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012c6f280 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012c6f380 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012c6f480 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012c6f580 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012c6f680 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012c6f780 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012c6f880 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200012cefbc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x2000192fdd00 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001967cec0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001967cfc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001967d0c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001967d1c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001967d2c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001967d3c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001967d4c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001967d5c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001967d6c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001967d7c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001967d8c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001967d9c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x2000196fdd00 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200019affc40 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200019defbc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200019defcc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200019ebc680 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4900c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4901c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4902c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4903c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4904c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4905c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4906c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4907c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4908c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4909c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b490ac0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b490bc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b490cc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b490dc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b490ec0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b490fc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4910c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4911c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4912c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4913c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4914c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4915c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4916c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4917c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4918c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4919c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b491ac0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b491bc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b491cc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b491dc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b491ec0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b491fc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4920c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4921c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4922c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4923c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4924c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4925c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4926c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4927c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4928c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4929c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b492ac0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b492bc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b492cc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b492dc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b492ec0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b492fc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4930c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4931c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4932c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4933c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4934c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4935c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4936c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4937c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4938c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4939c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b493ac0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b493bc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b493cc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b493dc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b493ec0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b493fc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4940c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4941c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4942c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4943c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4944c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4945c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4946c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4947c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4948c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4949c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b494ac0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b494bc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b494cc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b494dc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b494ec0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b494fc0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4950c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4951c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4952c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20001b4953c0 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200028864540 with size: 0.000244 MiB 00:07:09.217 element at address: 0x200028864640 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20002886b300 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20002886b580 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20002886b680 with size: 0.000244 MiB 00:07:09.217 element at address: 0x20002886b780 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886b880 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886b980 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886ba80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886bb80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886bc80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886bd80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886be80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886bf80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886c080 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886c180 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886c280 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886c380 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886c480 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886c580 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886c680 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886c780 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886c880 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886c980 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886ca80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886cb80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886cc80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886cd80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886ce80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886cf80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886d080 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886d180 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886d280 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886d380 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886d480 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886d580 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886d680 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886d780 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886d880 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886d980 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886da80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886db80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886dc80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886dd80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886de80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886df80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886e080 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886e180 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886e280 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886e380 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886e480 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886e580 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886e680 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886e780 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886e880 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886e980 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886ea80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886eb80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886ec80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886ed80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886ee80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886ef80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886f080 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886f180 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886f280 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886f380 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886f480 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886f580 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886f680 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886f780 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886f880 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886f980 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886fa80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886fb80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886fc80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886fd80 with size: 0.000244 MiB 00:07:09.218 element at address: 0x20002886fe80 with size: 0.000244 MiB 00:07:09.218 list of memzone associated elements. size: 607.930908 MiB 00:07:09.218 element at address: 0x20001b4954c0 with size: 211.416809 MiB 00:07:09.218 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:07:09.218 element at address: 0x20002886ff80 with size: 157.562622 MiB 00:07:09.218 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:07:09.218 element at address: 0x200012df1e40 with size: 100.055115 MiB 00:07:09.218 associated memzone info: size: 100.054932 MiB name: MP_bdev_io_57065_0 00:07:09.218 element at address: 0x200000dff340 with size: 48.003113 MiB 00:07:09.218 associated memzone info: size: 48.002930 MiB name: MP_msgpool_57065_0 00:07:09.218 element at address: 0x200003ffdb40 with size: 36.008972 MiB 00:07:09.218 associated memzone info: size: 36.008789 MiB name: MP_fsdev_io_57065_0 00:07:09.218 element at address: 0x200019fbe900 with size: 20.255615 MiB 00:07:09.218 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:07:09.218 element at address: 0x2000327feb00 with size: 18.005127 MiB 00:07:09.218 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:07:09.218 element at address: 0x2000004ffec0 with size: 3.000305 MiB 00:07:09.218 associated memzone info: size: 3.000122 MiB name: MP_evtpool_57065_0 00:07:09.218 element at address: 0x2000009ffdc0 with size: 2.000549 MiB 00:07:09.218 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_57065 00:07:09.218 element at address: 0x2000002d7c00 with size: 1.008179 MiB 00:07:09.218 associated memzone info: size: 1.007996 MiB name: MP_evtpool_57065 00:07:09.218 element at address: 0x2000196fde00 with size: 1.008179 MiB 00:07:09.218 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:07:09.218 element at address: 0x200019ebc780 with size: 1.008179 MiB 00:07:09.218 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:07:09.218 element at address: 0x2000192fde00 with size: 1.008179 MiB 00:07:09.218 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:07:09.218 element at address: 0x200012cefcc0 with size: 1.008179 MiB 00:07:09.218 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:07:09.218 element at address: 0x200000cff100 with size: 1.000549 MiB 00:07:09.218 associated memzone info: size: 1.000366 MiB name: RG_ring_0_57065 00:07:09.218 element at address: 0x2000008ffb80 with size: 1.000549 MiB 00:07:09.218 associated memzone info: size: 1.000366 MiB name: RG_ring_1_57065 00:07:09.218 element at address: 0x200019affd40 with size: 1.000549 MiB 00:07:09.218 associated memzone info: size: 1.000366 MiB name: RG_ring_4_57065 00:07:09.218 element at address: 0x2000326fe8c0 with size: 1.000549 MiB 00:07:09.218 associated memzone info: size: 1.000366 MiB name: RG_ring_5_57065 00:07:09.218 element at address: 0x20000087f5c0 with size: 0.500549 MiB 00:07:09.218 associated memzone info: size: 0.500366 MiB name: RG_MP_fsdev_io_57065 00:07:09.218 element at address: 0x200000c7ecc0 with size: 0.500549 MiB 00:07:09.218 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_57065 00:07:09.218 element at address: 0x20001967dac0 with size: 0.500549 MiB 00:07:09.218 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:07:09.218 element at address: 0x200012c6f980 with size: 0.500549 MiB 00:07:09.218 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:07:09.218 element at address: 0x200019e7c440 with size: 0.250549 MiB 00:07:09.218 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:07:09.218 element at address: 0x2000002b78c0 with size: 0.125549 MiB 00:07:09.218 associated memzone info: size: 0.125366 MiB name: RG_MP_evtpool_57065 00:07:09.218 element at address: 0x20000085df80 with size: 0.125549 MiB 00:07:09.218 associated memzone info: size: 0.125366 MiB name: RG_ring_2_57065 00:07:09.218 element at address: 0x2000192f5ac0 with size: 0.031799 MiB 00:07:09.218 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:07:09.218 element at address: 0x200028864740 with size: 0.023804 MiB 00:07:09.218 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:07:09.218 element at address: 0x200000859d40 with size: 0.016174 MiB 00:07:09.218 associated memzone info: size: 0.015991 MiB name: RG_ring_3_57065 00:07:09.218 element at address: 0x20002886a8c0 with size: 0.002502 MiB 00:07:09.218 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:07:09.218 element at address: 0x2000004ffa40 with size: 0.000366 MiB 00:07:09.218 associated memzone info: size: 0.000183 MiB name: MP_msgpool_57065 00:07:09.218 element at address: 0x2000008ff900 with size: 0.000366 MiB 00:07:09.218 associated memzone info: size: 0.000183 MiB name: MP_fsdev_io_57065 00:07:09.218 element at address: 0x200012bffd80 with size: 0.000366 MiB 00:07:09.218 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_57065 00:07:09.218 element at address: 0x20002886b400 with size: 0.000366 MiB 00:07:09.218 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:07:09.218 15:55:07 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:07:09.218 15:55:07 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 57065 00:07:09.218 15:55:07 dpdk_mem_utility -- common/autotest_common.sh@954 -- # '[' -z 57065 ']' 00:07:09.218 15:55:07 dpdk_mem_utility -- common/autotest_common.sh@958 -- # kill -0 57065 00:07:09.218 15:55:07 dpdk_mem_utility -- common/autotest_common.sh@959 -- # uname 00:07:09.218 15:55:07 dpdk_mem_utility -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:09.218 15:55:07 dpdk_mem_utility -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 57065 00:07:09.218 killing process with pid 57065 00:07:09.218 15:55:07 dpdk_mem_utility -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:09.218 15:55:07 dpdk_mem_utility -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:09.219 15:55:07 dpdk_mem_utility -- common/autotest_common.sh@972 -- # echo 'killing process with pid 57065' 00:07:09.219 15:55:07 dpdk_mem_utility -- common/autotest_common.sh@973 -- # kill 57065 00:07:09.219 15:55:07 dpdk_mem_utility -- common/autotest_common.sh@978 -- # wait 57065 00:07:11.123 00:07:11.123 real 0m2.744s 00:07:11.123 user 0m2.790s 00:07:11.123 sys 0m0.383s 00:07:11.123 15:55:08 dpdk_mem_utility -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:11.123 15:55:08 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:07:11.123 ************************************ 00:07:11.123 END TEST dpdk_mem_utility 00:07:11.123 ************************************ 00:07:11.123 15:55:08 -- spdk/autotest.sh@168 -- # run_test event /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:07:11.123 15:55:08 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:11.123 15:55:08 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:11.123 15:55:08 -- common/autotest_common.sh@10 -- # set +x 00:07:11.123 ************************************ 00:07:11.123 START TEST event 00:07:11.123 ************************************ 00:07:11.123 15:55:08 event -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:07:11.123 * Looking for test storage... 00:07:11.123 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:07:11.123 15:55:09 event -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:07:11.123 15:55:09 event -- common/autotest_common.sh@1693 -- # lcov --version 00:07:11.123 15:55:09 event -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:07:11.123 15:55:09 event -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:07:11.123 15:55:09 event -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:11.123 15:55:09 event -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:11.123 15:55:09 event -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:11.123 15:55:09 event -- scripts/common.sh@336 -- # IFS=.-: 00:07:11.123 15:55:09 event -- scripts/common.sh@336 -- # read -ra ver1 00:07:11.123 15:55:09 event -- scripts/common.sh@337 -- # IFS=.-: 00:07:11.123 15:55:09 event -- scripts/common.sh@337 -- # read -ra ver2 00:07:11.123 15:55:09 event -- scripts/common.sh@338 -- # local 'op=<' 00:07:11.123 15:55:09 event -- scripts/common.sh@340 -- # ver1_l=2 00:07:11.123 15:55:09 event -- scripts/common.sh@341 -- # ver2_l=1 00:07:11.123 15:55:09 event -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:11.123 15:55:09 event -- scripts/common.sh@344 -- # case "$op" in 00:07:11.123 15:55:09 event -- scripts/common.sh@345 -- # : 1 00:07:11.123 15:55:09 event -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:11.123 15:55:09 event -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:11.123 15:55:09 event -- scripts/common.sh@365 -- # decimal 1 00:07:11.123 15:55:09 event -- scripts/common.sh@353 -- # local d=1 00:07:11.123 15:55:09 event -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:11.123 15:55:09 event -- scripts/common.sh@355 -- # echo 1 00:07:11.123 15:55:09 event -- scripts/common.sh@365 -- # ver1[v]=1 00:07:11.123 15:55:09 event -- scripts/common.sh@366 -- # decimal 2 00:07:11.123 15:55:09 event -- scripts/common.sh@353 -- # local d=2 00:07:11.123 15:55:09 event -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:11.123 15:55:09 event -- scripts/common.sh@355 -- # echo 2 00:07:11.123 15:55:09 event -- scripts/common.sh@366 -- # ver2[v]=2 00:07:11.123 15:55:09 event -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:11.123 15:55:09 event -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:11.123 15:55:09 event -- scripts/common.sh@368 -- # return 0 00:07:11.123 15:55:09 event -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:11.123 15:55:09 event -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:07:11.123 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:11.123 --rc genhtml_branch_coverage=1 00:07:11.123 --rc genhtml_function_coverage=1 00:07:11.123 --rc genhtml_legend=1 00:07:11.123 --rc geninfo_all_blocks=1 00:07:11.123 --rc geninfo_unexecuted_blocks=1 00:07:11.123 00:07:11.123 ' 00:07:11.123 15:55:09 event -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:07:11.123 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:11.123 --rc genhtml_branch_coverage=1 00:07:11.123 --rc genhtml_function_coverage=1 00:07:11.123 --rc genhtml_legend=1 00:07:11.123 --rc geninfo_all_blocks=1 00:07:11.123 --rc geninfo_unexecuted_blocks=1 00:07:11.123 00:07:11.123 ' 00:07:11.123 15:55:09 event -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:07:11.123 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:11.123 --rc genhtml_branch_coverage=1 00:07:11.123 --rc genhtml_function_coverage=1 00:07:11.123 --rc genhtml_legend=1 00:07:11.123 --rc geninfo_all_blocks=1 00:07:11.123 --rc geninfo_unexecuted_blocks=1 00:07:11.123 00:07:11.123 ' 00:07:11.123 15:55:09 event -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:07:11.123 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:11.123 --rc genhtml_branch_coverage=1 00:07:11.123 --rc genhtml_function_coverage=1 00:07:11.123 --rc genhtml_legend=1 00:07:11.123 --rc geninfo_all_blocks=1 00:07:11.123 --rc geninfo_unexecuted_blocks=1 00:07:11.123 00:07:11.123 ' 00:07:11.123 15:55:09 event -- event/event.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:07:11.123 15:55:09 event -- bdev/nbd_common.sh@6 -- # set -e 00:07:11.123 15:55:09 event -- event/event.sh@45 -- # run_test event_perf /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:07:11.123 15:55:09 event -- common/autotest_common.sh@1105 -- # '[' 6 -le 1 ']' 00:07:11.123 15:55:09 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:11.123 15:55:09 event -- common/autotest_common.sh@10 -- # set +x 00:07:11.123 ************************************ 00:07:11.123 START TEST event_perf 00:07:11.123 ************************************ 00:07:11.123 15:55:09 event.event_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:07:11.123 Running I/O for 1 seconds...[2024-11-20 15:55:09.136824] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:07:11.123 [2024-11-20 15:55:09.136936] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57157 ] 00:07:11.123 [2024-11-20 15:55:09.296737] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:07:11.381 [2024-11-20 15:55:09.403100] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:11.381 [2024-11-20 15:55:09.403772] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:07:11.381 [2024-11-20 15:55:09.403969] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:11.381 Running I/O for 1 seconds...[2024-11-20 15:55:09.404307] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:07:12.317 00:07:12.317 lcore 0: 204732 00:07:12.317 lcore 1: 204728 00:07:12.317 lcore 2: 204728 00:07:12.317 lcore 3: 204731 00:07:12.317 done. 00:07:12.575 00:07:12.575 real 0m1.470s 00:07:12.575 user 0m4.262s 00:07:12.575 sys 0m0.086s 00:07:12.575 15:55:10 event.event_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:12.575 15:55:10 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:07:12.575 ************************************ 00:07:12.575 END TEST event_perf 00:07:12.575 ************************************ 00:07:12.575 15:55:10 event -- event/event.sh@46 -- # run_test event_reactor /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:07:12.576 15:55:10 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:07:12.576 15:55:10 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:12.576 15:55:10 event -- common/autotest_common.sh@10 -- # set +x 00:07:12.576 ************************************ 00:07:12.576 START TEST event_reactor 00:07:12.576 ************************************ 00:07:12.576 15:55:10 event.event_reactor -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:07:12.576 [2024-11-20 15:55:10.659522] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:07:12.576 [2024-11-20 15:55:10.659633] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57196 ] 00:07:12.576 [2024-11-20 15:55:10.815015] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:12.856 [2024-11-20 15:55:10.895220] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:13.793 test_start 00:07:13.793 oneshot 00:07:13.793 tick 100 00:07:13.793 tick 100 00:07:13.793 tick 250 00:07:13.793 tick 100 00:07:13.793 tick 100 00:07:13.793 tick 100 00:07:13.793 tick 250 00:07:13.793 tick 500 00:07:13.793 tick 100 00:07:13.793 tick 100 00:07:13.793 tick 250 00:07:13.793 tick 100 00:07:13.793 tick 100 00:07:13.793 test_end 00:07:13.793 00:07:13.793 real 0m1.391s 00:07:13.793 user 0m1.224s 00:07:13.793 sys 0m0.058s 00:07:13.793 ************************************ 00:07:13.793 END TEST event_reactor 00:07:13.793 ************************************ 00:07:13.793 15:55:12 event.event_reactor -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:13.793 15:55:12 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:07:14.052 15:55:12 event -- event/event.sh@47 -- # run_test event_reactor_perf /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:07:14.052 15:55:12 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:07:14.052 15:55:12 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:14.052 15:55:12 event -- common/autotest_common.sh@10 -- # set +x 00:07:14.052 ************************************ 00:07:14.052 START TEST event_reactor_perf 00:07:14.052 ************************************ 00:07:14.052 15:55:12 event.event_reactor_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:07:14.052 [2024-11-20 15:55:12.111502] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:07:14.052 [2024-11-20 15:55:12.111638] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57233 ] 00:07:14.052 [2024-11-20 15:55:12.262759] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:14.310 [2024-11-20 15:55:12.345776] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:15.244 test_start 00:07:15.244 test_end 00:07:15.244 Performance: 401570 events per second 00:07:15.244 00:07:15.244 real 0m1.389s 00:07:15.244 user 0m1.225s 00:07:15.244 sys 0m0.058s 00:07:15.244 15:55:13 event.event_reactor_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:15.244 15:55:13 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:07:15.244 ************************************ 00:07:15.244 END TEST event_reactor_perf 00:07:15.244 ************************************ 00:07:15.510 15:55:13 event -- event/event.sh@49 -- # uname -s 00:07:15.510 15:55:13 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:07:15.510 15:55:13 event -- event/event.sh@50 -- # run_test event_scheduler /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:07:15.510 15:55:13 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:15.510 15:55:13 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:15.510 15:55:13 event -- common/autotest_common.sh@10 -- # set +x 00:07:15.510 ************************************ 00:07:15.510 START TEST event_scheduler 00:07:15.510 ************************************ 00:07:15.510 15:55:13 event.event_scheduler -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:07:15.510 * Looking for test storage... 00:07:15.510 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event/scheduler 00:07:15.510 15:55:13 event.event_scheduler -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:07:15.510 15:55:13 event.event_scheduler -- common/autotest_common.sh@1693 -- # lcov --version 00:07:15.510 15:55:13 event.event_scheduler -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:07:15.510 15:55:13 event.event_scheduler -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@336 -- # IFS=.-: 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@336 -- # read -ra ver1 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@337 -- # IFS=.-: 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@337 -- # read -ra ver2 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@338 -- # local 'op=<' 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@340 -- # ver1_l=2 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@341 -- # ver2_l=1 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@344 -- # case "$op" in 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@345 -- # : 1 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@365 -- # decimal 1 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@353 -- # local d=1 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@355 -- # echo 1 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@365 -- # ver1[v]=1 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@366 -- # decimal 2 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@353 -- # local d=2 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@355 -- # echo 2 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@366 -- # ver2[v]=2 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:15.510 15:55:13 event.event_scheduler -- scripts/common.sh@368 -- # return 0 00:07:15.510 15:55:13 event.event_scheduler -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:15.510 15:55:13 event.event_scheduler -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:07:15.510 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:15.510 --rc genhtml_branch_coverage=1 00:07:15.510 --rc genhtml_function_coverage=1 00:07:15.510 --rc genhtml_legend=1 00:07:15.510 --rc geninfo_all_blocks=1 00:07:15.510 --rc geninfo_unexecuted_blocks=1 00:07:15.510 00:07:15.510 ' 00:07:15.510 15:55:13 event.event_scheduler -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:07:15.510 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:15.510 --rc genhtml_branch_coverage=1 00:07:15.510 --rc genhtml_function_coverage=1 00:07:15.510 --rc genhtml_legend=1 00:07:15.510 --rc geninfo_all_blocks=1 00:07:15.510 --rc geninfo_unexecuted_blocks=1 00:07:15.510 00:07:15.510 ' 00:07:15.510 15:55:13 event.event_scheduler -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:07:15.510 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:15.510 --rc genhtml_branch_coverage=1 00:07:15.510 --rc genhtml_function_coverage=1 00:07:15.510 --rc genhtml_legend=1 00:07:15.510 --rc geninfo_all_blocks=1 00:07:15.510 --rc geninfo_unexecuted_blocks=1 00:07:15.510 00:07:15.510 ' 00:07:15.510 15:55:13 event.event_scheduler -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:07:15.510 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:15.510 --rc genhtml_branch_coverage=1 00:07:15.510 --rc genhtml_function_coverage=1 00:07:15.510 --rc genhtml_legend=1 00:07:15.510 --rc geninfo_all_blocks=1 00:07:15.510 --rc geninfo_unexecuted_blocks=1 00:07:15.510 00:07:15.510 ' 00:07:15.510 15:55:13 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:07:15.510 15:55:13 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=57303 00:07:15.510 15:55:13 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:07:15.510 15:55:13 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 57303 00:07:15.510 15:55:13 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:07:15.510 15:55:13 event.event_scheduler -- common/autotest_common.sh@835 -- # '[' -z 57303 ']' 00:07:15.510 15:55:13 event.event_scheduler -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:15.510 15:55:13 event.event_scheduler -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:15.510 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:15.510 15:55:13 event.event_scheduler -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:15.510 15:55:13 event.event_scheduler -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:15.510 15:55:13 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:07:15.510 [2024-11-20 15:55:13.753479] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:07:15.510 [2024-11-20 15:55:13.753602] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57303 ] 00:07:15.770 [2024-11-20 15:55:13.912263] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:07:16.096 [2024-11-20 15:55:14.021887] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:16.096 [2024-11-20 15:55:14.022200] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:16.096 [2024-11-20 15:55:14.023212] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:07:16.096 [2024-11-20 15:55:14.023298] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:07:16.671 15:55:14 event.event_scheduler -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:16.671 15:55:14 event.event_scheduler -- common/autotest_common.sh@868 -- # return 0 00:07:16.671 15:55:14 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:07:16.671 15:55:14 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.671 15:55:14 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:07:16.671 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:07:16.671 POWER: Cannot set governor of lcore 0 to userspace 00:07:16.671 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:07:16.671 POWER: Cannot set governor of lcore 0 to performance 00:07:16.671 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:07:16.671 POWER: Cannot set governor of lcore 0 to userspace 00:07:16.671 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:07:16.671 POWER: Cannot set governor of lcore 0 to userspace 00:07:16.671 GUEST_CHANNEL: Opening channel '/dev/virtio-ports/virtio.serial.port.poweragent.0' for lcore 0 00:07:16.671 GUEST_CHANNEL: Unable to connect to '/dev/virtio-ports/virtio.serial.port.poweragent.0' with error No such file or directory 00:07:16.671 POWER: Unable to set Power Management Environment for lcore 0 00:07:16.671 [2024-11-20 15:55:14.636721] dpdk_governor.c: 135:_init_core: *ERROR*: Failed to initialize on core0 00:07:16.671 [2024-11-20 15:55:14.636742] dpdk_governor.c: 196:_init: *ERROR*: Failed to initialize on core0 00:07:16.671 [2024-11-20 15:55:14.636752] scheduler_dynamic.c: 280:init: *NOTICE*: Unable to initialize dpdk governor 00:07:16.671 [2024-11-20 15:55:14.636771] scheduler_dynamic.c: 427:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:07:16.671 [2024-11-20 15:55:14.636779] scheduler_dynamic.c: 429:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:07:16.671 [2024-11-20 15:55:14.636788] scheduler_dynamic.c: 431:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:07:16.671 15:55:14 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.671 15:55:14 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:07:16.671 15:55:14 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.671 15:55:14 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:07:16.671 [2024-11-20 15:55:14.868171] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:07:16.671 15:55:14 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.671 15:55:14 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:07:16.671 15:55:14 event.event_scheduler -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:16.671 15:55:14 event.event_scheduler -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:16.671 15:55:14 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:07:16.671 ************************************ 00:07:16.671 START TEST scheduler_create_thread 00:07:16.671 ************************************ 00:07:16.671 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1129 -- # scheduler_create_thread 00:07:16.671 15:55:14 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:07:16.671 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.671 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:16.671 2 00:07:16.671 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.671 15:55:14 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:07:16.671 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.671 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:16.671 3 00:07:16.671 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.671 15:55:14 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:07:16.671 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.671 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:16.671 4 00:07:16.671 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.671 15:55:14 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:07:16.671 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.671 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:16.932 5 00:07:16.932 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.932 15:55:14 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:07:16.932 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:16.933 6 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:16.933 7 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:16.933 8 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:16.933 9 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:16.933 10 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:16.933 15:55:14 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:17.872 15:55:16 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.872 00:07:17.872 real 0m1.175s 00:07:17.872 user 0m0.014s 00:07:17.872 sys 0m0.006s 00:07:17.872 15:55:16 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:17.872 15:55:16 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:17.872 ************************************ 00:07:17.872 END TEST scheduler_create_thread 00:07:17.872 ************************************ 00:07:17.872 15:55:16 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:07:17.872 15:55:16 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 57303 00:07:17.872 15:55:16 event.event_scheduler -- common/autotest_common.sh@954 -- # '[' -z 57303 ']' 00:07:17.872 15:55:16 event.event_scheduler -- common/autotest_common.sh@958 -- # kill -0 57303 00:07:17.872 15:55:16 event.event_scheduler -- common/autotest_common.sh@959 -- # uname 00:07:17.872 15:55:16 event.event_scheduler -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:18.132 15:55:16 event.event_scheduler -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 57303 00:07:18.132 15:55:16 event.event_scheduler -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:07:18.132 killing process with pid 57303 00:07:18.132 15:55:16 event.event_scheduler -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:07:18.132 15:55:16 event.event_scheduler -- common/autotest_common.sh@972 -- # echo 'killing process with pid 57303' 00:07:18.132 15:55:16 event.event_scheduler -- common/autotest_common.sh@973 -- # kill 57303 00:07:18.132 15:55:16 event.event_scheduler -- common/autotest_common.sh@978 -- # wait 57303 00:07:18.392 [2024-11-20 15:55:16.541575] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:07:19.406 00:07:19.406 real 0m3.767s 00:07:19.406 user 0m6.232s 00:07:19.406 sys 0m0.368s 00:07:19.406 15:55:17 event.event_scheduler -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:19.406 ************************************ 00:07:19.406 END TEST event_scheduler 00:07:19.406 15:55:17 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:07:19.406 ************************************ 00:07:19.406 15:55:17 event -- event/event.sh@51 -- # modprobe -n nbd 00:07:19.406 15:55:17 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:07:19.406 15:55:17 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:19.406 15:55:17 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:19.406 15:55:17 event -- common/autotest_common.sh@10 -- # set +x 00:07:19.406 ************************************ 00:07:19.406 START TEST app_repeat 00:07:19.406 ************************************ 00:07:19.406 15:55:17 event.app_repeat -- common/autotest_common.sh@1129 -- # app_repeat_test 00:07:19.406 15:55:17 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:19.406 15:55:17 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:19.406 15:55:17 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:07:19.406 15:55:17 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:19.406 15:55:17 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:07:19.406 15:55:17 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:07:19.406 15:55:17 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:07:19.406 Process app_repeat pid: 57393 00:07:19.406 15:55:17 event.app_repeat -- event/event.sh@19 -- # repeat_pid=57393 00:07:19.406 15:55:17 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:07:19.406 15:55:17 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 57393' 00:07:19.406 15:55:17 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:07:19.406 spdk_app_start Round 0 00:07:19.406 15:55:17 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:07:19.406 15:55:17 event.app_repeat -- event/event.sh@25 -- # waitforlisten 57393 /var/tmp/spdk-nbd.sock 00:07:19.406 15:55:17 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 57393 ']' 00:07:19.406 15:55:17 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:07:19.406 15:55:17 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:19.406 15:55:17 event.app_repeat -- event/event.sh@18 -- # /home/vagrant/spdk_repo/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:07:19.406 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:07:19.406 15:55:17 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:07:19.406 15:55:17 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:19.406 15:55:17 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:07:19.406 [2024-11-20 15:55:17.402540] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:07:19.406 [2024-11-20 15:55:17.402656] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57393 ] 00:07:19.406 [2024-11-20 15:55:17.561108] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:07:19.666 [2024-11-20 15:55:17.666579] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:19.666 [2024-11-20 15:55:17.666776] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:20.238 15:55:18 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:20.238 15:55:18 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:07:20.238 15:55:18 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:20.498 Malloc0 00:07:20.498 15:55:18 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:20.758 Malloc1 00:07:20.758 15:55:18 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:20.758 15:55:18 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:20.758 15:55:18 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:20.758 15:55:18 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:07:20.758 15:55:18 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:20.758 15:55:18 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:07:20.758 15:55:18 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:20.758 15:55:18 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:20.758 15:55:18 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:20.758 15:55:18 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:07:20.758 15:55:18 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:20.758 15:55:18 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:07:20.758 15:55:18 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:07:20.758 15:55:18 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:07:20.758 15:55:18 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:20.758 15:55:18 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:07:20.758 /dev/nbd0 00:07:21.018 15:55:19 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:07:21.018 15:55:19 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:07:21.018 15:55:19 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:07:21.018 15:55:19 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:07:21.018 15:55:19 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:07:21.018 15:55:19 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:07:21.018 15:55:19 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:07:21.018 15:55:19 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:07:21.018 15:55:19 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:07:21.018 15:55:19 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:07:21.018 15:55:19 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:21.018 1+0 records in 00:07:21.018 1+0 records out 00:07:21.018 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000396683 s, 10.3 MB/s 00:07:21.018 15:55:19 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:21.018 15:55:19 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:07:21.018 15:55:19 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:21.018 15:55:19 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:07:21.018 15:55:19 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:07:21.018 15:55:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:21.018 15:55:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:21.018 15:55:19 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:07:21.018 /dev/nbd1 00:07:21.276 15:55:19 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:07:21.276 15:55:19 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:07:21.276 15:55:19 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:07:21.276 15:55:19 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:07:21.276 15:55:19 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:07:21.276 15:55:19 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:07:21.276 15:55:19 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:07:21.276 15:55:19 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:07:21.276 15:55:19 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:07:21.276 15:55:19 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:07:21.276 15:55:19 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:21.276 1+0 records in 00:07:21.276 1+0 records out 00:07:21.276 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000309652 s, 13.2 MB/s 00:07:21.276 15:55:19 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:21.276 15:55:19 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:07:21.276 15:55:19 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:21.276 15:55:19 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:07:21.276 15:55:19 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:07:21.276 15:55:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:21.276 15:55:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:21.276 15:55:19 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:21.276 15:55:19 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:07:21.277 { 00:07:21.277 "nbd_device": "/dev/nbd0", 00:07:21.277 "bdev_name": "Malloc0" 00:07:21.277 }, 00:07:21.277 { 00:07:21.277 "nbd_device": "/dev/nbd1", 00:07:21.277 "bdev_name": "Malloc1" 00:07:21.277 } 00:07:21.277 ]' 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:07:21.277 { 00:07:21.277 "nbd_device": "/dev/nbd0", 00:07:21.277 "bdev_name": "Malloc0" 00:07:21.277 }, 00:07:21.277 { 00:07:21.277 "nbd_device": "/dev/nbd1", 00:07:21.277 "bdev_name": "Malloc1" 00:07:21.277 } 00:07:21.277 ]' 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:07:21.277 /dev/nbd1' 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:07:21.277 /dev/nbd1' 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:07:21.277 15:55:19 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:07:21.534 256+0 records in 00:07:21.534 256+0 records out 00:07:21.534 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0111369 s, 94.2 MB/s 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:07:21.534 256+0 records in 00:07:21.534 256+0 records out 00:07:21.534 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0226621 s, 46.3 MB/s 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:07:21.534 256+0 records in 00:07:21.534 256+0 records out 00:07:21.534 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0271793 s, 38.6 MB/s 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:21.534 15:55:19 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:07:21.792 15:55:19 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:07:21.792 15:55:19 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:07:21.792 15:55:19 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:07:21.792 15:55:19 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:21.792 15:55:19 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:21.792 15:55:19 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:07:21.792 15:55:19 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:21.792 15:55:19 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:21.792 15:55:19 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:21.792 15:55:19 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:07:21.792 15:55:20 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:07:22.049 15:55:20 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:07:22.049 15:55:20 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:07:22.049 15:55:20 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:22.049 15:55:20 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:22.049 15:55:20 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:07:22.049 15:55:20 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:22.049 15:55:20 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:22.049 15:55:20 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:22.049 15:55:20 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:22.049 15:55:20 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:22.049 15:55:20 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:07:22.049 15:55:20 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:07:22.049 15:55:20 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:22.049 15:55:20 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:07:22.049 15:55:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:22.308 15:55:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:07:22.308 15:55:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:07:22.308 15:55:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:07:22.308 15:55:20 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:07:22.308 15:55:20 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:07:22.308 15:55:20 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:07:22.308 15:55:20 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:07:22.308 15:55:20 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:07:22.565 15:55:20 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:07:23.508 [2024-11-20 15:55:21.464877] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:07:23.508 [2024-11-20 15:55:21.566208] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:23.508 [2024-11-20 15:55:21.566356] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:23.508 [2024-11-20 15:55:21.688977] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:07:23.508 [2024-11-20 15:55:21.689053] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:07:25.436 15:55:23 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:07:25.436 spdk_app_start Round 1 00:07:25.436 15:55:23 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:07:25.436 15:55:23 event.app_repeat -- event/event.sh@25 -- # waitforlisten 57393 /var/tmp/spdk-nbd.sock 00:07:25.436 15:55:23 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 57393 ']' 00:07:25.436 15:55:23 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:07:25.436 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:07:25.436 15:55:23 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:25.436 15:55:23 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:07:25.436 15:55:23 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:25.436 15:55:23 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:07:25.696 15:55:23 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:25.696 15:55:23 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:07:25.696 15:55:23 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:25.958 Malloc0 00:07:25.958 15:55:24 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:26.218 Malloc1 00:07:26.218 15:55:24 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:26.218 15:55:24 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:26.218 15:55:24 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:26.218 15:55:24 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:07:26.218 15:55:24 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:26.218 15:55:24 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:07:26.218 15:55:24 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:26.218 15:55:24 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:26.218 15:55:24 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:26.218 15:55:24 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:07:26.218 15:55:24 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:26.218 15:55:24 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:07:26.218 15:55:24 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:07:26.218 15:55:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:07:26.218 15:55:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:26.218 15:55:24 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:07:26.479 /dev/nbd0 00:07:26.479 15:55:24 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:07:26.479 15:55:24 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:07:26.479 15:55:24 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:07:26.479 15:55:24 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:07:26.479 15:55:24 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:07:26.479 15:55:24 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:07:26.479 15:55:24 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:07:26.479 15:55:24 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:07:26.479 15:55:24 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:07:26.479 15:55:24 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:07:26.479 15:55:24 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:26.479 1+0 records in 00:07:26.479 1+0 records out 00:07:26.479 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000215075 s, 19.0 MB/s 00:07:26.479 15:55:24 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:26.479 15:55:24 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:07:26.479 15:55:24 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:26.479 15:55:24 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:07:26.479 15:55:24 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:07:26.479 15:55:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:26.479 15:55:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:26.479 15:55:24 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:07:26.740 /dev/nbd1 00:07:26.740 15:55:24 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:07:26.740 15:55:24 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:07:26.740 15:55:24 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:07:26.740 15:55:24 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:07:26.740 15:55:24 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:07:26.740 15:55:24 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:07:26.740 15:55:24 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:07:26.740 15:55:24 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:07:26.740 15:55:24 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:07:26.740 15:55:24 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:07:26.740 15:55:24 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:26.740 1+0 records in 00:07:26.740 1+0 records out 00:07:26.740 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000195483 s, 21.0 MB/s 00:07:26.740 15:55:24 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:26.740 15:55:24 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:07:26.740 15:55:24 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:26.740 15:55:24 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:07:26.740 15:55:24 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:07:26.740 15:55:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:26.740 15:55:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:26.740 15:55:24 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:26.740 15:55:24 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:26.740 15:55:24 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:07:27.000 { 00:07:27.000 "nbd_device": "/dev/nbd0", 00:07:27.000 "bdev_name": "Malloc0" 00:07:27.000 }, 00:07:27.000 { 00:07:27.000 "nbd_device": "/dev/nbd1", 00:07:27.000 "bdev_name": "Malloc1" 00:07:27.000 } 00:07:27.000 ]' 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:07:27.000 { 00:07:27.000 "nbd_device": "/dev/nbd0", 00:07:27.000 "bdev_name": "Malloc0" 00:07:27.000 }, 00:07:27.000 { 00:07:27.000 "nbd_device": "/dev/nbd1", 00:07:27.000 "bdev_name": "Malloc1" 00:07:27.000 } 00:07:27.000 ]' 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:07:27.000 /dev/nbd1' 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:07:27.000 /dev/nbd1' 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:07:27.000 256+0 records in 00:07:27.000 256+0 records out 00:07:27.000 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00807885 s, 130 MB/s 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:07:27.000 256+0 records in 00:07:27.000 256+0 records out 00:07:27.000 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.02114 s, 49.6 MB/s 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:07:27.000 256+0 records in 00:07:27.000 256+0 records out 00:07:27.000 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0231529 s, 45.3 MB/s 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:27.000 15:55:25 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:07:27.260 15:55:25 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:07:27.260 15:55:25 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:07:27.260 15:55:25 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:07:27.260 15:55:25 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:27.260 15:55:25 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:27.260 15:55:25 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:07:27.260 15:55:25 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:27.260 15:55:25 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:27.260 15:55:25 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:27.260 15:55:25 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:07:27.522 15:55:25 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:07:27.522 15:55:25 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:07:27.522 15:55:25 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:07:27.522 15:55:25 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:27.522 15:55:25 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:27.522 15:55:25 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:07:27.522 15:55:25 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:27.522 15:55:25 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:27.522 15:55:25 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:27.522 15:55:25 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:27.522 15:55:25 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:27.782 15:55:25 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:07:27.782 15:55:25 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:07:27.782 15:55:25 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:27.782 15:55:25 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:07:27.782 15:55:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:07:27.782 15:55:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:27.782 15:55:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:07:27.782 15:55:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:07:27.782 15:55:25 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:07:27.782 15:55:25 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:07:27.782 15:55:25 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:07:27.782 15:55:25 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:07:27.782 15:55:25 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:07:28.083 15:55:26 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:07:28.657 [2024-11-20 15:55:26.886579] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:07:28.917 [2024-11-20 15:55:26.986484] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:28.917 [2024-11-20 15:55:26.986628] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:28.917 [2024-11-20 15:55:27.115849] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:07:28.917 [2024-11-20 15:55:27.115915] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:07:31.454 spdk_app_start Round 2 00:07:31.454 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:07:31.454 15:55:29 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:07:31.454 15:55:29 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:07:31.454 15:55:29 event.app_repeat -- event/event.sh@25 -- # waitforlisten 57393 /var/tmp/spdk-nbd.sock 00:07:31.454 15:55:29 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 57393 ']' 00:07:31.454 15:55:29 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:07:31.454 15:55:29 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:31.454 15:55:29 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:07:31.454 15:55:29 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:31.454 15:55:29 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:07:31.454 15:55:29 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:31.454 15:55:29 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:07:31.454 15:55:29 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:31.454 Malloc0 00:07:31.454 15:55:29 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:31.714 Malloc1 00:07:31.714 15:55:29 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:31.714 15:55:29 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:31.714 15:55:29 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:31.714 15:55:29 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:07:31.714 15:55:29 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:31.714 15:55:29 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:07:31.714 15:55:29 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:31.714 15:55:29 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:31.714 15:55:29 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:31.714 15:55:29 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:07:31.714 15:55:29 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:31.714 15:55:29 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:07:31.714 15:55:29 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:07:31.714 15:55:29 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:07:31.714 15:55:29 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:31.714 15:55:29 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:07:31.975 /dev/nbd0 00:07:31.975 15:55:30 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:07:31.975 15:55:30 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:07:31.975 15:55:30 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:07:31.975 15:55:30 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:07:31.975 15:55:30 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:07:31.975 15:55:30 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:07:31.975 15:55:30 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:07:31.975 15:55:30 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:07:31.975 15:55:30 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:07:31.975 15:55:30 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:07:31.975 15:55:30 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:31.975 1+0 records in 00:07:31.975 1+0 records out 00:07:31.975 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000258351 s, 15.9 MB/s 00:07:31.975 15:55:30 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:31.975 15:55:30 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:07:31.975 15:55:30 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:31.975 15:55:30 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:07:31.975 15:55:30 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:07:31.975 15:55:30 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:31.975 15:55:30 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:31.975 15:55:30 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:07:32.237 /dev/nbd1 00:07:32.237 15:55:30 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:07:32.237 15:55:30 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:07:32.237 15:55:30 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:07:32.237 15:55:30 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:07:32.237 15:55:30 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:07:32.237 15:55:30 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:07:32.237 15:55:30 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:07:32.237 15:55:30 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:07:32.237 15:55:30 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:07:32.237 15:55:30 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:07:32.237 15:55:30 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:32.237 1+0 records in 00:07:32.237 1+0 records out 00:07:32.237 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000217772 s, 18.8 MB/s 00:07:32.237 15:55:30 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:32.237 15:55:30 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:07:32.237 15:55:30 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:32.237 15:55:30 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:07:32.237 15:55:30 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:07:32.237 15:55:30 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:32.237 15:55:30 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:32.237 15:55:30 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:32.237 15:55:30 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:32.237 15:55:30 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:32.496 15:55:30 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:07:32.496 { 00:07:32.496 "nbd_device": "/dev/nbd0", 00:07:32.496 "bdev_name": "Malloc0" 00:07:32.496 }, 00:07:32.496 { 00:07:32.497 "nbd_device": "/dev/nbd1", 00:07:32.497 "bdev_name": "Malloc1" 00:07:32.497 } 00:07:32.497 ]' 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:07:32.497 { 00:07:32.497 "nbd_device": "/dev/nbd0", 00:07:32.497 "bdev_name": "Malloc0" 00:07:32.497 }, 00:07:32.497 { 00:07:32.497 "nbd_device": "/dev/nbd1", 00:07:32.497 "bdev_name": "Malloc1" 00:07:32.497 } 00:07:32.497 ]' 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:07:32.497 /dev/nbd1' 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:07:32.497 /dev/nbd1' 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:07:32.497 256+0 records in 00:07:32.497 256+0 records out 00:07:32.497 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00751038 s, 140 MB/s 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:07:32.497 256+0 records in 00:07:32.497 256+0 records out 00:07:32.497 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0208495 s, 50.3 MB/s 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:07:32.497 256+0 records in 00:07:32.497 256+0 records out 00:07:32.497 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0229648 s, 45.7 MB/s 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:32.497 15:55:30 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:33.069 15:55:31 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:33.329 15:55:31 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:07:33.329 15:55:31 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:33.329 15:55:31 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:07:33.329 15:55:31 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:07:33.329 15:55:31 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:33.329 15:55:31 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:07:33.329 15:55:31 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:07:33.329 15:55:31 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:07:33.329 15:55:31 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:07:33.329 15:55:31 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:07:33.329 15:55:31 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:07:33.329 15:55:31 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:07:33.329 15:55:31 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:07:33.900 15:55:31 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:07:34.472 [2024-11-20 15:55:32.561564] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:07:34.472 [2024-11-20 15:55:32.659735] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:34.472 [2024-11-20 15:55:32.659746] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:34.734 [2024-11-20 15:55:32.786040] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:07:34.734 [2024-11-20 15:55:32.786100] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:07:36.644 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:07:36.644 15:55:34 event.app_repeat -- event/event.sh@38 -- # waitforlisten 57393 /var/tmp/spdk-nbd.sock 00:07:36.644 15:55:34 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 57393 ']' 00:07:36.644 15:55:34 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:07:36.644 15:55:34 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:36.644 15:55:34 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:07:36.644 15:55:34 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:36.644 15:55:34 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:07:36.906 15:55:35 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:36.906 15:55:35 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:07:36.906 15:55:35 event.app_repeat -- event/event.sh@39 -- # killprocess 57393 00:07:36.906 15:55:35 event.app_repeat -- common/autotest_common.sh@954 -- # '[' -z 57393 ']' 00:07:36.906 15:55:35 event.app_repeat -- common/autotest_common.sh@958 -- # kill -0 57393 00:07:36.906 15:55:35 event.app_repeat -- common/autotest_common.sh@959 -- # uname 00:07:36.906 15:55:35 event.app_repeat -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:36.906 15:55:35 event.app_repeat -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 57393 00:07:36.906 killing process with pid 57393 00:07:36.906 15:55:35 event.app_repeat -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:36.906 15:55:35 event.app_repeat -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:36.906 15:55:35 event.app_repeat -- common/autotest_common.sh@972 -- # echo 'killing process with pid 57393' 00:07:36.906 15:55:35 event.app_repeat -- common/autotest_common.sh@973 -- # kill 57393 00:07:36.906 15:55:35 event.app_repeat -- common/autotest_common.sh@978 -- # wait 57393 00:07:37.847 spdk_app_start is called in Round 0. 00:07:37.847 Shutdown signal received, stop current app iteration 00:07:37.847 Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 reinitialization... 00:07:37.847 spdk_app_start is called in Round 1. 00:07:37.847 Shutdown signal received, stop current app iteration 00:07:37.847 Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 reinitialization... 00:07:37.847 spdk_app_start is called in Round 2. 00:07:37.847 Shutdown signal received, stop current app iteration 00:07:37.847 Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 reinitialization... 00:07:37.847 spdk_app_start is called in Round 3. 00:07:37.847 Shutdown signal received, stop current app iteration 00:07:37.847 15:55:35 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:07:37.847 15:55:35 event.app_repeat -- event/event.sh@42 -- # return 0 00:07:37.847 00:07:37.847 real 0m18.399s 00:07:37.847 user 0m40.126s 00:07:37.847 sys 0m2.271s 00:07:37.847 15:55:35 event.app_repeat -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:37.847 15:55:35 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:07:37.847 ************************************ 00:07:37.847 END TEST app_repeat 00:07:37.847 ************************************ 00:07:37.847 15:55:35 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:07:37.847 15:55:35 event -- event/event.sh@55 -- # run_test cpu_locks /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:07:37.847 15:55:35 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:37.847 15:55:35 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:37.847 15:55:35 event -- common/autotest_common.sh@10 -- # set +x 00:07:37.847 ************************************ 00:07:37.847 START TEST cpu_locks 00:07:37.847 ************************************ 00:07:37.847 15:55:35 event.cpu_locks -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:07:37.847 * Looking for test storage... 00:07:37.847 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:07:37.847 15:55:35 event.cpu_locks -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:07:37.847 15:55:35 event.cpu_locks -- common/autotest_common.sh@1693 -- # lcov --version 00:07:37.847 15:55:35 event.cpu_locks -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:07:37.847 15:55:35 event.cpu_locks -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@336 -- # IFS=.-: 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@336 -- # read -ra ver1 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@337 -- # IFS=.-: 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@337 -- # read -ra ver2 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@338 -- # local 'op=<' 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@340 -- # ver1_l=2 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@341 -- # ver2_l=1 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@344 -- # case "$op" in 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@345 -- # : 1 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@365 -- # decimal 1 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@353 -- # local d=1 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@355 -- # echo 1 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@365 -- # ver1[v]=1 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@366 -- # decimal 2 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@353 -- # local d=2 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@355 -- # echo 2 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@366 -- # ver2[v]=2 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:37.847 15:55:35 event.cpu_locks -- scripts/common.sh@368 -- # return 0 00:07:37.847 15:55:35 event.cpu_locks -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:37.847 15:55:35 event.cpu_locks -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:07:37.847 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:37.847 --rc genhtml_branch_coverage=1 00:07:37.847 --rc genhtml_function_coverage=1 00:07:37.847 --rc genhtml_legend=1 00:07:37.847 --rc geninfo_all_blocks=1 00:07:37.847 --rc geninfo_unexecuted_blocks=1 00:07:37.847 00:07:37.847 ' 00:07:37.847 15:55:35 event.cpu_locks -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:07:37.847 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:37.847 --rc genhtml_branch_coverage=1 00:07:37.847 --rc genhtml_function_coverage=1 00:07:37.847 --rc genhtml_legend=1 00:07:37.847 --rc geninfo_all_blocks=1 00:07:37.847 --rc geninfo_unexecuted_blocks=1 00:07:37.847 00:07:37.847 ' 00:07:37.847 15:55:35 event.cpu_locks -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:07:37.847 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:37.847 --rc genhtml_branch_coverage=1 00:07:37.847 --rc genhtml_function_coverage=1 00:07:37.847 --rc genhtml_legend=1 00:07:37.847 --rc geninfo_all_blocks=1 00:07:37.847 --rc geninfo_unexecuted_blocks=1 00:07:37.847 00:07:37.847 ' 00:07:37.847 15:55:35 event.cpu_locks -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:07:37.847 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:37.847 --rc genhtml_branch_coverage=1 00:07:37.847 --rc genhtml_function_coverage=1 00:07:37.847 --rc genhtml_legend=1 00:07:37.847 --rc geninfo_all_blocks=1 00:07:37.847 --rc geninfo_unexecuted_blocks=1 00:07:37.847 00:07:37.847 ' 00:07:37.847 15:55:35 event.cpu_locks -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:07:37.847 15:55:35 event.cpu_locks -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:07:37.847 15:55:35 event.cpu_locks -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:07:37.847 15:55:35 event.cpu_locks -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:07:37.847 15:55:35 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:37.847 15:55:35 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:37.847 15:55:35 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:37.847 ************************************ 00:07:37.847 START TEST default_locks 00:07:37.847 ************************************ 00:07:37.847 15:55:35 event.cpu_locks.default_locks -- common/autotest_common.sh@1129 -- # default_locks 00:07:37.847 15:55:36 event.cpu_locks.default_locks -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=57829 00:07:37.847 15:55:36 event.cpu_locks.default_locks -- event/cpu_locks.sh@47 -- # waitforlisten 57829 00:07:37.847 15:55:36 event.cpu_locks.default_locks -- event/cpu_locks.sh@45 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:07:37.847 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:37.847 15:55:36 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 57829 ']' 00:07:37.847 15:55:36 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:37.847 15:55:36 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:37.847 15:55:36 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:37.847 15:55:36 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:37.847 15:55:36 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:07:38.108 [2024-11-20 15:55:36.102393] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:07:38.108 [2024-11-20 15:55:36.102560] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57829 ] 00:07:38.108 [2024-11-20 15:55:36.278301] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:38.367 [2024-11-20 15:55:36.381559] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:38.936 15:55:36 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:38.936 15:55:36 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 0 00:07:38.936 15:55:36 event.cpu_locks.default_locks -- event/cpu_locks.sh@49 -- # locks_exist 57829 00:07:38.936 15:55:36 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:07:38.936 15:55:36 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # lslocks -p 57829 00:07:38.936 15:55:37 event.cpu_locks.default_locks -- event/cpu_locks.sh@50 -- # killprocess 57829 00:07:38.936 15:55:37 event.cpu_locks.default_locks -- common/autotest_common.sh@954 -- # '[' -z 57829 ']' 00:07:38.936 15:55:37 event.cpu_locks.default_locks -- common/autotest_common.sh@958 -- # kill -0 57829 00:07:38.936 15:55:37 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # uname 00:07:38.936 15:55:37 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:38.936 15:55:37 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 57829 00:07:38.936 15:55:37 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:38.936 15:55:37 event.cpu_locks.default_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:38.936 killing process with pid 57829 00:07:38.936 15:55:37 event.cpu_locks.default_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 57829' 00:07:38.936 15:55:37 event.cpu_locks.default_locks -- common/autotest_common.sh@973 -- # kill 57829 00:07:38.936 15:55:37 event.cpu_locks.default_locks -- common/autotest_common.sh@978 -- # wait 57829 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- event/cpu_locks.sh@52 -- # NOT waitforlisten 57829 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@652 -- # local es=0 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 57829 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # waitforlisten 57829 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 57829 ']' 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:40.873 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:07:40.873 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (57829) - No such process 00:07:40.873 ERROR: process (pid: 57829) is no longer running 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 1 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # es=1 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:07:40.873 ************************************ 00:07:40.873 END TEST default_locks 00:07:40.873 ************************************ 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- event/cpu_locks.sh@54 -- # no_locks 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # lock_files=() 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # local lock_files 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:07:40.873 00:07:40.873 real 0m2.712s 00:07:40.873 user 0m2.684s 00:07:40.873 sys 0m0.465s 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:40.873 15:55:38 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:07:40.873 15:55:38 event.cpu_locks -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:07:40.873 15:55:38 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:40.873 15:55:38 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:40.873 15:55:38 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:40.873 ************************************ 00:07:40.873 START TEST default_locks_via_rpc 00:07:40.873 ************************************ 00:07:40.873 15:55:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1129 -- # default_locks_via_rpc 00:07:40.873 15:55:38 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=57893 00:07:40.873 15:55:38 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:07:40.873 15:55:38 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@63 -- # waitforlisten 57893 00:07:40.873 15:55:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 57893 ']' 00:07:40.873 15:55:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:40.873 15:55:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:40.873 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:40.873 15:55:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:40.873 15:55:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:40.873 15:55:38 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:40.873 [2024-11-20 15:55:38.885491] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:07:40.873 [2024-11-20 15:55:38.885678] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57893 ] 00:07:40.873 [2024-11-20 15:55:39.062762] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:41.133 [2024-11-20 15:55:39.167031] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:41.707 15:55:39 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:41.707 15:55:39 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:07:41.707 15:55:39 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:07:41.707 15:55:39 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.707 15:55:39 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:41.707 15:55:39 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.707 15:55:39 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@67 -- # no_locks 00:07:41.707 15:55:39 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # lock_files=() 00:07:41.707 15:55:39 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # local lock_files 00:07:41.707 15:55:39 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:07:41.707 15:55:39 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:07:41.707 15:55:39 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.707 15:55:39 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:41.707 15:55:39 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.707 15:55:39 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@71 -- # locks_exist 57893 00:07:41.707 15:55:39 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # lslocks -p 57893 00:07:41.707 15:55:39 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:07:41.968 15:55:39 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@73 -- # killprocess 57893 00:07:41.968 15:55:39 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@954 -- # '[' -z 57893 ']' 00:07:41.968 15:55:39 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@958 -- # kill -0 57893 00:07:41.968 15:55:39 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # uname 00:07:41.968 15:55:39 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:41.968 15:55:39 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 57893 00:07:41.968 15:55:40 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:41.968 15:55:40 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:41.968 15:55:40 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 57893' 00:07:41.968 killing process with pid 57893 00:07:41.968 15:55:40 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@973 -- # kill 57893 00:07:41.968 15:55:40 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@978 -- # wait 57893 00:07:43.351 00:07:43.351 real 0m2.757s 00:07:43.351 user 0m2.714s 00:07:43.351 sys 0m0.494s 00:07:43.351 15:55:41 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:43.351 ************************************ 00:07:43.351 END TEST default_locks_via_rpc 00:07:43.351 ************************************ 00:07:43.351 15:55:41 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:43.351 15:55:41 event.cpu_locks -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:07:43.351 15:55:41 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:43.351 15:55:41 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:43.351 15:55:41 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:43.351 ************************************ 00:07:43.351 START TEST non_locking_app_on_locked_coremask 00:07:43.351 ************************************ 00:07:43.351 15:55:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # non_locking_app_on_locked_coremask 00:07:43.351 15:55:41 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=57956 00:07:43.351 15:55:41 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@81 -- # waitforlisten 57956 /var/tmp/spdk.sock 00:07:43.351 15:55:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 57956 ']' 00:07:43.351 15:55:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:43.351 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:43.351 15:55:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:43.351 15:55:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:43.351 15:55:41 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@79 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:07:43.352 15:55:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:43.352 15:55:41 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:43.612 [2024-11-20 15:55:41.670163] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:07:43.612 [2024-11-20 15:55:41.670282] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57956 ] 00:07:43.612 [2024-11-20 15:55:41.829312] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:43.873 [2024-11-20 15:55:41.931474] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:44.443 15:55:42 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:44.443 15:55:42 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:07:44.443 15:55:42 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=57972 00:07:44.443 15:55:42 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@85 -- # waitforlisten 57972 /var/tmp/spdk2.sock 00:07:44.443 15:55:42 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@83 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:07:44.443 15:55:42 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 57972 ']' 00:07:44.444 15:55:42 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:07:44.444 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:07:44.444 15:55:42 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:44.444 15:55:42 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:07:44.444 15:55:42 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:44.444 15:55:42 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:44.444 [2024-11-20 15:55:42.593574] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:07:44.444 [2024-11-20 15:55:42.593711] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57972 ] 00:07:44.704 [2024-11-20 15:55:42.764934] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:07:44.704 [2024-11-20 15:55:42.764992] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:44.963 [2024-11-20 15:55:42.971901] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:45.924 15:55:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:45.924 15:55:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:07:45.924 15:55:44 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@87 -- # locks_exist 57956 00:07:45.924 15:55:44 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 57956 00:07:45.924 15:55:44 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:07:46.494 15:55:44 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@89 -- # killprocess 57956 00:07:46.494 15:55:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 57956 ']' 00:07:46.494 15:55:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 57956 00:07:46.494 15:55:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:07:46.494 15:55:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:46.494 15:55:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 57956 00:07:46.494 15:55:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:46.494 15:55:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:46.494 killing process with pid 57956 00:07:46.494 15:55:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 57956' 00:07:46.494 15:55:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 57956 00:07:46.494 15:55:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 57956 00:07:49.832 15:55:47 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@90 -- # killprocess 57972 00:07:49.832 15:55:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 57972 ']' 00:07:49.832 15:55:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 57972 00:07:49.832 15:55:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:07:49.832 15:55:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:49.832 15:55:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 57972 00:07:49.832 15:55:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:49.832 killing process with pid 57972 00:07:49.832 15:55:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:49.832 15:55:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 57972' 00:07:49.832 15:55:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 57972 00:07:49.832 15:55:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 57972 00:07:51.218 00:07:51.218 real 0m7.493s 00:07:51.218 user 0m7.729s 00:07:51.218 sys 0m0.893s 00:07:51.218 15:55:49 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:51.218 15:55:49 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:51.218 ************************************ 00:07:51.218 END TEST non_locking_app_on_locked_coremask 00:07:51.218 ************************************ 00:07:51.218 15:55:49 event.cpu_locks -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:07:51.218 15:55:49 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:51.218 15:55:49 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:51.218 15:55:49 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:51.218 ************************************ 00:07:51.218 START TEST locking_app_on_unlocked_coremask 00:07:51.218 ************************************ 00:07:51.218 15:55:49 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_unlocked_coremask 00:07:51.218 15:55:49 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=58074 00:07:51.218 15:55:49 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@99 -- # waitforlisten 58074 /var/tmp/spdk.sock 00:07:51.218 15:55:49 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 58074 ']' 00:07:51.218 15:55:49 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:51.218 15:55:49 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:51.218 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:51.218 15:55:49 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:51.218 15:55:49 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@97 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:07:51.218 15:55:49 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:51.218 15:55:49 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:51.218 [2024-11-20 15:55:49.214963] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:07:51.218 [2024-11-20 15:55:49.215242] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58074 ] 00:07:51.218 [2024-11-20 15:55:49.373080] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:07:51.218 [2024-11-20 15:55:49.373133] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:51.479 [2024-11-20 15:55:49.475576] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:52.049 15:55:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:52.049 15:55:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:07:52.049 15:55:50 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=58090 00:07:52.049 15:55:50 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@103 -- # waitforlisten 58090 /var/tmp/spdk2.sock 00:07:52.049 15:55:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 58090 ']' 00:07:52.049 15:55:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:07:52.049 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:07:52.049 15:55:50 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@101 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:07:52.049 15:55:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:52.049 15:55:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:07:52.049 15:55:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:52.049 15:55:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:52.049 [2024-11-20 15:55:50.142929] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:07:52.049 [2024-11-20 15:55:50.143054] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58090 ] 00:07:52.310 [2024-11-20 15:55:50.317952] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:52.310 [2024-11-20 15:55:50.525184] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:53.693 15:55:51 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:53.693 15:55:51 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:07:53.693 15:55:51 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@105 -- # locks_exist 58090 00:07:53.693 15:55:51 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:07:53.693 15:55:51 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 58090 00:07:53.954 15:55:52 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@107 -- # killprocess 58074 00:07:53.954 15:55:52 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 58074 ']' 00:07:53.954 15:55:52 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 58074 00:07:53.954 15:55:52 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:07:53.954 15:55:52 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:53.954 15:55:52 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58074 00:07:53.954 killing process with pid 58074 00:07:53.954 15:55:52 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:53.954 15:55:52 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:53.954 15:55:52 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58074' 00:07:53.954 15:55:52 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 58074 00:07:53.954 15:55:52 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 58074 00:07:57.255 15:55:55 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@108 -- # killprocess 58090 00:07:57.255 15:55:55 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 58090 ']' 00:07:57.255 15:55:55 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 58090 00:07:57.255 15:55:55 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:07:57.255 15:55:55 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:57.255 15:55:55 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58090 00:07:57.255 killing process with pid 58090 00:07:57.255 15:55:55 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:57.255 15:55:55 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:57.255 15:55:55 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58090' 00:07:57.255 15:55:55 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 58090 00:07:57.255 15:55:55 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 58090 00:07:58.674 00:07:58.674 real 0m7.589s 00:07:58.674 user 0m7.831s 00:07:58.674 sys 0m0.850s 00:07:58.674 ************************************ 00:07:58.674 END TEST locking_app_on_unlocked_coremask 00:07:58.674 ************************************ 00:07:58.674 15:55:56 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:58.674 15:55:56 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:58.674 15:55:56 event.cpu_locks -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:07:58.674 15:55:56 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:58.674 15:55:56 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:58.674 15:55:56 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:58.674 ************************************ 00:07:58.674 START TEST locking_app_on_locked_coremask 00:07:58.674 ************************************ 00:07:58.674 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:58.674 15:55:56 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_locked_coremask 00:07:58.674 15:55:56 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=58198 00:07:58.674 15:55:56 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@116 -- # waitforlisten 58198 /var/tmp/spdk.sock 00:07:58.674 15:55:56 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 58198 ']' 00:07:58.674 15:55:56 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:58.674 15:55:56 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@114 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:07:58.674 15:55:56 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:58.674 15:55:56 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:58.674 15:55:56 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:58.674 15:55:56 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:58.674 [2024-11-20 15:55:56.870743] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:07:58.674 [2024-11-20 15:55:56.870868] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58198 ] 00:07:58.936 [2024-11-20 15:55:57.023282] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:58.936 [2024-11-20 15:55:57.127756] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@118 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=58214 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@120 -- # NOT waitforlisten 58214 /var/tmp/spdk2.sock 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@652 -- # local es=0 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 58214 /var/tmp/spdk2.sock 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:07:59.509 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # waitforlisten 58214 /var/tmp/spdk2.sock 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 58214 ']' 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:59.509 15:55:57 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:59.770 [2024-11-20 15:55:57.800727] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:07:59.770 [2024-11-20 15:55:57.800879] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58214 ] 00:07:59.770 [2024-11-20 15:55:57.976006] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 58198 has claimed it. 00:07:59.770 [2024-11-20 15:55:57.976071] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:08:00.340 ERROR: process (pid: 58214) is no longer running 00:08:00.340 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (58214) - No such process 00:08:00.340 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:00.340 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 1 00:08:00.340 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # es=1 00:08:00.340 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:00.340 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:00.340 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:00.340 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@122 -- # locks_exist 58198 00:08:00.340 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 58198 00:08:00.340 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:08:00.600 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@124 -- # killprocess 58198 00:08:00.600 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 58198 ']' 00:08:00.600 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 58198 00:08:00.600 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:08:00.600 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:00.600 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58198 00:08:00.600 killing process with pid 58198 00:08:00.600 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:00.600 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:00.600 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58198' 00:08:00.600 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 58198 00:08:00.600 15:55:58 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 58198 00:08:02.559 00:08:02.559 real 0m3.496s 00:08:02.559 user 0m3.699s 00:08:02.559 sys 0m0.596s 00:08:02.559 15:56:00 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:02.559 ************************************ 00:08:02.559 END TEST locking_app_on_locked_coremask 00:08:02.559 ************************************ 00:08:02.559 15:56:00 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:08:02.559 15:56:00 event.cpu_locks -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:08:02.559 15:56:00 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:02.559 15:56:00 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:02.559 15:56:00 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:08:02.559 ************************************ 00:08:02.559 START TEST locking_overlapped_coremask 00:08:02.559 ************************************ 00:08:02.559 15:56:00 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask 00:08:02.559 15:56:00 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=58272 00:08:02.559 15:56:00 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@133 -- # waitforlisten 58272 /var/tmp/spdk.sock 00:08:02.559 15:56:00 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 58272 ']' 00:08:02.559 15:56:00 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:02.559 15:56:00 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:02.559 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:02.559 15:56:00 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:02.559 15:56:00 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:02.559 15:56:00 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:08:02.559 15:56:00 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@131 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 00:08:02.559 [2024-11-20 15:56:00.426889] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:02.559 [2024-11-20 15:56:00.427478] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58272 ] 00:08:02.559 [2024-11-20 15:56:00.584740] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:08:02.559 [2024-11-20 15:56:00.691956] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:08:02.559 [2024-11-20 15:56:00.692316] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:08:02.559 [2024-11-20 15:56:00.692421] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 0 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=58290 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@137 -- # NOT waitforlisten 58290 /var/tmp/spdk2.sock 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@652 -- # local es=0 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 58290 /var/tmp/spdk2.sock 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@135 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # waitforlisten 58290 /var/tmp/spdk2.sock 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 58290 ']' 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:08:03.131 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:03.131 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:08:03.131 [2024-11-20 15:56:01.371263] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:03.131 [2024-11-20 15:56:01.371533] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58290 ] 00:08:03.393 [2024-11-20 15:56:01.546802] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 58272 has claimed it. 00:08:03.393 [2024-11-20 15:56:01.546863] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:08:03.962 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (58290) - No such process 00:08:03.962 ERROR: process (pid: 58290) is no longer running 00:08:03.962 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:03.962 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 1 00:08:03.962 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # es=1 00:08:03.962 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:03.962 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:03.962 15:56:01 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:03.962 15:56:01 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:08:03.962 15:56:02 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:08:03.962 15:56:02 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:08:03.962 15:56:02 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:08:03.962 15:56:02 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@141 -- # killprocess 58272 00:08:03.962 15:56:02 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@954 -- # '[' -z 58272 ']' 00:08:03.962 15:56:02 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@958 -- # kill -0 58272 00:08:03.962 15:56:02 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # uname 00:08:03.962 15:56:02 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:03.962 15:56:02 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58272 00:08:03.962 15:56:02 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:03.962 15:56:02 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:03.962 15:56:02 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58272' 00:08:03.962 killing process with pid 58272 00:08:03.962 15:56:02 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@973 -- # kill 58272 00:08:03.963 15:56:02 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@978 -- # wait 58272 00:08:05.407 00:08:05.407 real 0m3.227s 00:08:05.407 user 0m8.723s 00:08:05.407 sys 0m0.471s 00:08:05.408 15:56:03 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:05.408 ************************************ 00:08:05.408 15:56:03 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:08:05.408 END TEST locking_overlapped_coremask 00:08:05.408 ************************************ 00:08:05.408 15:56:03 event.cpu_locks -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:08:05.408 15:56:03 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:05.408 15:56:03 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:05.408 15:56:03 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:08:05.408 ************************************ 00:08:05.408 START TEST locking_overlapped_coremask_via_rpc 00:08:05.408 ************************************ 00:08:05.408 15:56:03 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask_via_rpc 00:08:05.408 15:56:03 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=58343 00:08:05.408 15:56:03 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@149 -- # waitforlisten 58343 /var/tmp/spdk.sock 00:08:05.408 15:56:03 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@147 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:08:05.408 15:56:03 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 58343 ']' 00:08:05.408 15:56:03 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:05.408 15:56:03 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:05.408 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:05.408 15:56:03 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:05.408 15:56:03 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:05.408 15:56:03 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:05.668 [2024-11-20 15:56:03.716787] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:05.668 [2024-11-20 15:56:03.716905] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58343 ] 00:08:05.668 [2024-11-20 15:56:03.876616] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:08:05.668 [2024-11-20 15:56:03.876662] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:08:05.928 [2024-11-20 15:56:03.981862] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:08:05.928 [2024-11-20 15:56:03.982231] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:08:05.928 [2024-11-20 15:56:03.982445] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:06.498 15:56:04 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:06.498 15:56:04 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:08:06.498 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:08:06.498 15:56:04 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=58361 00:08:06.498 15:56:04 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@153 -- # waitforlisten 58361 /var/tmp/spdk2.sock 00:08:06.498 15:56:04 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 58361 ']' 00:08:06.498 15:56:04 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:08:06.498 15:56:04 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@151 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:08:06.498 15:56:04 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:06.498 15:56:04 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:08:06.498 15:56:04 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:06.498 15:56:04 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:06.498 [2024-11-20 15:56:04.661699] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:06.498 [2024-11-20 15:56:04.662385] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58361 ] 00:08:06.759 [2024-11-20 15:56:04.836619] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:08:06.759 [2024-11-20 15:56:04.836676] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:08:07.020 [2024-11-20 15:56:05.050952] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:08:07.020 [2024-11-20 15:56:05.051121] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:08:07.020 [2024-11-20 15:56:05.051143] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:08:08.404 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:08.404 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:08:08.404 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:08:08.404 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.404 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:08.404 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.404 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:08:08.404 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@652 -- # local es=0 00:08:08.404 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:08:08.404 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:08:08.404 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:08.404 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:08:08.404 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:08.404 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:08:08.404 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.404 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:08.404 [2024-11-20 15:56:06.253827] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 58343 has claimed it. 00:08:08.404 request: 00:08:08.404 { 00:08:08.404 "method": "framework_enable_cpumask_locks", 00:08:08.404 "req_id": 1 00:08:08.404 } 00:08:08.404 Got JSON-RPC error response 00:08:08.404 response: 00:08:08.404 { 00:08:08.404 "code": -32603, 00:08:08.404 "message": "Failed to claim CPU core: 2" 00:08:08.404 } 00:08:08.404 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:08.404 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # es=1 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@158 -- # waitforlisten 58343 /var/tmp/spdk.sock 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 58343 ']' 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:08.405 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@159 -- # waitforlisten 58361 /var/tmp/spdk2.sock 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 58361 ']' 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:08.405 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:08.667 ************************************ 00:08:08.667 END TEST locking_overlapped_coremask_via_rpc 00:08:08.667 ************************************ 00:08:08.667 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:08.667 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:08:08.667 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:08:08.667 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:08:08.667 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:08:08.667 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:08:08.667 00:08:08.667 real 0m3.045s 00:08:08.667 user 0m1.099s 00:08:08.667 sys 0m0.121s 00:08:08.667 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:08.667 15:56:06 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:08.667 15:56:06 event.cpu_locks -- event/cpu_locks.sh@174 -- # cleanup 00:08:08.667 15:56:06 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 58343 ]] 00:08:08.667 15:56:06 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 58343 00:08:08.667 15:56:06 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 58343 ']' 00:08:08.667 15:56:06 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 58343 00:08:08.667 15:56:06 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:08:08.667 15:56:06 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:08.667 15:56:06 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58343 00:08:08.667 killing process with pid 58343 00:08:08.667 15:56:06 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:08.667 15:56:06 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:08.667 15:56:06 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58343' 00:08:08.667 15:56:06 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 58343 00:08:08.667 15:56:06 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 58343 00:08:10.053 15:56:08 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 58361 ]] 00:08:10.053 15:56:08 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 58361 00:08:10.053 15:56:08 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 58361 ']' 00:08:10.053 15:56:08 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 58361 00:08:10.053 15:56:08 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:08:10.313 15:56:08 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:10.313 15:56:08 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58361 00:08:10.313 killing process with pid 58361 00:08:10.313 15:56:08 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:08:10.313 15:56:08 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:08:10.313 15:56:08 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58361' 00:08:10.313 15:56:08 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 58361 00:08:10.313 15:56:08 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 58361 00:08:11.794 15:56:09 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:08:11.794 15:56:09 event.cpu_locks -- event/cpu_locks.sh@1 -- # cleanup 00:08:11.794 15:56:09 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 58343 ]] 00:08:11.794 Process with pid 58343 is not found 00:08:11.794 Process with pid 58361 is not found 00:08:11.794 15:56:09 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 58343 00:08:11.794 15:56:09 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 58343 ']' 00:08:11.794 15:56:09 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 58343 00:08:11.794 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (58343) - No such process 00:08:11.794 15:56:09 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 58343 is not found' 00:08:11.794 15:56:09 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 58361 ]] 00:08:11.794 15:56:09 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 58361 00:08:11.794 15:56:09 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 58361 ']' 00:08:11.794 15:56:09 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 58361 00:08:11.794 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (58361) - No such process 00:08:11.794 15:56:09 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 58361 is not found' 00:08:11.794 15:56:09 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:08:11.795 00:08:11.795 real 0m34.057s 00:08:11.795 user 0m57.911s 00:08:11.795 sys 0m4.748s 00:08:11.795 15:56:09 event.cpu_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:11.795 15:56:09 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:08:11.795 ************************************ 00:08:11.795 END TEST cpu_locks 00:08:11.795 ************************************ 00:08:11.795 00:08:11.795 real 1m0.998s 00:08:11.795 user 1m51.155s 00:08:11.795 sys 0m7.829s 00:08:11.795 15:56:09 event -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:11.795 ************************************ 00:08:11.795 END TEST event 00:08:11.795 15:56:09 event -- common/autotest_common.sh@10 -- # set +x 00:08:11.795 ************************************ 00:08:11.795 15:56:09 -- spdk/autotest.sh@169 -- # run_test thread /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:08:11.795 15:56:09 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:11.795 15:56:09 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:11.795 15:56:09 -- common/autotest_common.sh@10 -- # set +x 00:08:11.795 ************************************ 00:08:11.795 START TEST thread 00:08:11.795 ************************************ 00:08:11.795 15:56:09 thread -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:08:12.055 * Looking for test storage... 00:08:12.055 * Found test storage at /home/vagrant/spdk_repo/spdk/test/thread 00:08:12.055 15:56:10 thread -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:08:12.055 15:56:10 thread -- common/autotest_common.sh@1693 -- # lcov --version 00:08:12.055 15:56:10 thread -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:08:12.055 15:56:10 thread -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:08:12.055 15:56:10 thread -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:12.055 15:56:10 thread -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:12.055 15:56:10 thread -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:12.055 15:56:10 thread -- scripts/common.sh@336 -- # IFS=.-: 00:08:12.055 15:56:10 thread -- scripts/common.sh@336 -- # read -ra ver1 00:08:12.055 15:56:10 thread -- scripts/common.sh@337 -- # IFS=.-: 00:08:12.055 15:56:10 thread -- scripts/common.sh@337 -- # read -ra ver2 00:08:12.055 15:56:10 thread -- scripts/common.sh@338 -- # local 'op=<' 00:08:12.055 15:56:10 thread -- scripts/common.sh@340 -- # ver1_l=2 00:08:12.055 15:56:10 thread -- scripts/common.sh@341 -- # ver2_l=1 00:08:12.055 15:56:10 thread -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:12.055 15:56:10 thread -- scripts/common.sh@344 -- # case "$op" in 00:08:12.055 15:56:10 thread -- scripts/common.sh@345 -- # : 1 00:08:12.055 15:56:10 thread -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:12.055 15:56:10 thread -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:12.055 15:56:10 thread -- scripts/common.sh@365 -- # decimal 1 00:08:12.055 15:56:10 thread -- scripts/common.sh@353 -- # local d=1 00:08:12.055 15:56:10 thread -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:12.055 15:56:10 thread -- scripts/common.sh@355 -- # echo 1 00:08:12.056 15:56:10 thread -- scripts/common.sh@365 -- # ver1[v]=1 00:08:12.056 15:56:10 thread -- scripts/common.sh@366 -- # decimal 2 00:08:12.056 15:56:10 thread -- scripts/common.sh@353 -- # local d=2 00:08:12.056 15:56:10 thread -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:12.056 15:56:10 thread -- scripts/common.sh@355 -- # echo 2 00:08:12.056 15:56:10 thread -- scripts/common.sh@366 -- # ver2[v]=2 00:08:12.056 15:56:10 thread -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:12.056 15:56:10 thread -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:12.056 15:56:10 thread -- scripts/common.sh@368 -- # return 0 00:08:12.056 15:56:10 thread -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:12.056 15:56:10 thread -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:08:12.056 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:12.056 --rc genhtml_branch_coverage=1 00:08:12.056 --rc genhtml_function_coverage=1 00:08:12.056 --rc genhtml_legend=1 00:08:12.056 --rc geninfo_all_blocks=1 00:08:12.056 --rc geninfo_unexecuted_blocks=1 00:08:12.056 00:08:12.056 ' 00:08:12.056 15:56:10 thread -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:08:12.056 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:12.056 --rc genhtml_branch_coverage=1 00:08:12.056 --rc genhtml_function_coverage=1 00:08:12.056 --rc genhtml_legend=1 00:08:12.056 --rc geninfo_all_blocks=1 00:08:12.056 --rc geninfo_unexecuted_blocks=1 00:08:12.056 00:08:12.056 ' 00:08:12.056 15:56:10 thread -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:08:12.056 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:12.056 --rc genhtml_branch_coverage=1 00:08:12.056 --rc genhtml_function_coverage=1 00:08:12.056 --rc genhtml_legend=1 00:08:12.056 --rc geninfo_all_blocks=1 00:08:12.056 --rc geninfo_unexecuted_blocks=1 00:08:12.056 00:08:12.056 ' 00:08:12.056 15:56:10 thread -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:08:12.056 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:12.056 --rc genhtml_branch_coverage=1 00:08:12.056 --rc genhtml_function_coverage=1 00:08:12.056 --rc genhtml_legend=1 00:08:12.056 --rc geninfo_all_blocks=1 00:08:12.056 --rc geninfo_unexecuted_blocks=1 00:08:12.056 00:08:12.056 ' 00:08:12.056 15:56:10 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:08:12.056 15:56:10 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:08:12.056 15:56:10 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:12.056 15:56:10 thread -- common/autotest_common.sh@10 -- # set +x 00:08:12.056 ************************************ 00:08:12.056 START TEST thread_poller_perf 00:08:12.056 ************************************ 00:08:12.056 15:56:10 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:08:12.056 [2024-11-20 15:56:10.174360] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:12.056 [2024-11-20 15:56:10.174468] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58527 ] 00:08:12.317 [2024-11-20 15:56:10.335632] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:12.317 [2024-11-20 15:56:10.439191] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:12.317 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:08:13.702 [2024-11-20T15:56:11.953Z] ====================================== 00:08:13.702 [2024-11-20T15:56:11.953Z] busy:2613446124 (cyc) 00:08:13.702 [2024-11-20T15:56:11.953Z] total_run_count: 302000 00:08:13.702 [2024-11-20T15:56:11.953Z] tsc_hz: 2600000000 (cyc) 00:08:13.702 [2024-11-20T15:56:11.953Z] ====================================== 00:08:13.702 [2024-11-20T15:56:11.953Z] poller_cost: 8653 (cyc), 3328 (nsec) 00:08:13.702 00:08:13.702 real 0m1.462s 00:08:13.702 user 0m1.287s 00:08:13.702 sys 0m0.066s 00:08:13.702 15:56:11 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:13.702 ************************************ 00:08:13.702 END TEST thread_poller_perf 00:08:13.702 15:56:11 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:08:13.702 ************************************ 00:08:13.702 15:56:11 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:08:13.702 15:56:11 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:08:13.702 15:56:11 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:13.702 15:56:11 thread -- common/autotest_common.sh@10 -- # set +x 00:08:13.702 ************************************ 00:08:13.702 START TEST thread_poller_perf 00:08:13.702 ************************************ 00:08:13.702 15:56:11 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:08:13.702 [2024-11-20 15:56:11.692240] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:13.702 [2024-11-20 15:56:11.692523] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58563 ] 00:08:13.702 [2024-11-20 15:56:11.850880] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:13.961 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:08:13.961 [2024-11-20 15:56:11.959803] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:14.901 [2024-11-20T15:56:13.152Z] ====================================== 00:08:14.901 [2024-11-20T15:56:13.152Z] busy:2603343072 (cyc) 00:08:14.901 [2024-11-20T15:56:13.152Z] total_run_count: 3912000 00:08:14.901 [2024-11-20T15:56:13.152Z] tsc_hz: 2600000000 (cyc) 00:08:14.902 [2024-11-20T15:56:13.153Z] ====================================== 00:08:14.902 [2024-11-20T15:56:13.153Z] poller_cost: 665 (cyc), 255 (nsec) 00:08:14.902 00:08:14.902 real 0m1.457s 00:08:14.902 user 0m1.280s 00:08:14.902 sys 0m0.068s 00:08:14.902 ************************************ 00:08:14.902 END TEST thread_poller_perf 00:08:14.902 ************************************ 00:08:14.902 15:56:13 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:14.902 15:56:13 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:08:15.162 15:56:13 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:08:15.162 ************************************ 00:08:15.162 END TEST thread 00:08:15.162 ************************************ 00:08:15.162 00:08:15.162 real 0m3.177s 00:08:15.162 user 0m2.682s 00:08:15.162 sys 0m0.239s 00:08:15.162 15:56:13 thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:15.162 15:56:13 thread -- common/autotest_common.sh@10 -- # set +x 00:08:15.162 15:56:13 -- spdk/autotest.sh@171 -- # [[ 0 -eq 1 ]] 00:08:15.162 15:56:13 -- spdk/autotest.sh@176 -- # run_test app_cmdline /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:08:15.162 15:56:13 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:15.162 15:56:13 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:15.162 15:56:13 -- common/autotest_common.sh@10 -- # set +x 00:08:15.162 ************************************ 00:08:15.162 START TEST app_cmdline 00:08:15.162 ************************************ 00:08:15.162 15:56:13 app_cmdline -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:08:15.162 * Looking for test storage... 00:08:15.162 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:08:15.162 15:56:13 app_cmdline -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:08:15.162 15:56:13 app_cmdline -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:08:15.162 15:56:13 app_cmdline -- common/autotest_common.sh@1693 -- # lcov --version 00:08:15.162 15:56:13 app_cmdline -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@336 -- # IFS=.-: 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@336 -- # read -ra ver1 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@337 -- # IFS=.-: 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@337 -- # read -ra ver2 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@338 -- # local 'op=<' 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@340 -- # ver1_l=2 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@341 -- # ver2_l=1 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@344 -- # case "$op" in 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@345 -- # : 1 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@365 -- # decimal 1 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@353 -- # local d=1 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@355 -- # echo 1 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@365 -- # ver1[v]=1 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@366 -- # decimal 2 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@353 -- # local d=2 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@355 -- # echo 2 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@366 -- # ver2[v]=2 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:15.162 15:56:13 app_cmdline -- scripts/common.sh@368 -- # return 0 00:08:15.162 15:56:13 app_cmdline -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:15.162 15:56:13 app_cmdline -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:08:15.162 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:15.162 --rc genhtml_branch_coverage=1 00:08:15.162 --rc genhtml_function_coverage=1 00:08:15.162 --rc genhtml_legend=1 00:08:15.162 --rc geninfo_all_blocks=1 00:08:15.162 --rc geninfo_unexecuted_blocks=1 00:08:15.162 00:08:15.162 ' 00:08:15.162 15:56:13 app_cmdline -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:08:15.162 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:15.162 --rc genhtml_branch_coverage=1 00:08:15.162 --rc genhtml_function_coverage=1 00:08:15.162 --rc genhtml_legend=1 00:08:15.162 --rc geninfo_all_blocks=1 00:08:15.162 --rc geninfo_unexecuted_blocks=1 00:08:15.162 00:08:15.162 ' 00:08:15.162 15:56:13 app_cmdline -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:08:15.162 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:15.162 --rc genhtml_branch_coverage=1 00:08:15.162 --rc genhtml_function_coverage=1 00:08:15.162 --rc genhtml_legend=1 00:08:15.162 --rc geninfo_all_blocks=1 00:08:15.162 --rc geninfo_unexecuted_blocks=1 00:08:15.162 00:08:15.162 ' 00:08:15.162 15:56:13 app_cmdline -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:08:15.162 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:15.162 --rc genhtml_branch_coverage=1 00:08:15.162 --rc genhtml_function_coverage=1 00:08:15.162 --rc genhtml_legend=1 00:08:15.162 --rc geninfo_all_blocks=1 00:08:15.162 --rc geninfo_unexecuted_blocks=1 00:08:15.162 00:08:15.162 ' 00:08:15.162 15:56:13 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:08:15.162 15:56:13 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=58647 00:08:15.162 15:56:13 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 58647 00:08:15.162 15:56:13 app_cmdline -- common/autotest_common.sh@835 -- # '[' -z 58647 ']' 00:08:15.162 15:56:13 app_cmdline -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:15.162 15:56:13 app_cmdline -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:15.162 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:15.162 15:56:13 app_cmdline -- app/cmdline.sh@16 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:08:15.163 15:56:13 app_cmdline -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:15.163 15:56:13 app_cmdline -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:15.163 15:56:13 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:08:15.423 [2024-11-20 15:56:13.449729] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:15.423 [2024-11-20 15:56:13.449853] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58647 ] 00:08:15.423 [2024-11-20 15:56:13.601598] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:15.685 [2024-11-20 15:56:13.704591] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:16.256 15:56:14 app_cmdline -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:16.256 15:56:14 app_cmdline -- common/autotest_common.sh@868 -- # return 0 00:08:16.256 15:56:14 app_cmdline -- app/cmdline.sh@20 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py spdk_get_version 00:08:16.256 { 00:08:16.256 "version": "SPDK v25.01-pre git sha1 0728de5b0", 00:08:16.256 "fields": { 00:08:16.256 "major": 25, 00:08:16.256 "minor": 1, 00:08:16.256 "patch": 0, 00:08:16.256 "suffix": "-pre", 00:08:16.256 "commit": "0728de5b0" 00:08:16.256 } 00:08:16.256 } 00:08:16.517 15:56:14 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:08:16.517 15:56:14 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:08:16.517 15:56:14 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:08:16.517 15:56:14 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:08:16.517 15:56:14 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:08:16.517 15:56:14 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:08:16.517 15:56:14 app_cmdline -- app/cmdline.sh@26 -- # sort 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.517 15:56:14 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:08:16.517 15:56:14 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:08:16.517 15:56:14 app_cmdline -- app/cmdline.sh@30 -- # NOT /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@652 -- # local es=0 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/scripts/rpc.py ]] 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:08:16.517 request: 00:08:16.517 { 00:08:16.517 "method": "env_dpdk_get_mem_stats", 00:08:16.517 "req_id": 1 00:08:16.517 } 00:08:16.517 Got JSON-RPC error response 00:08:16.517 response: 00:08:16.517 { 00:08:16.517 "code": -32601, 00:08:16.517 "message": "Method not found" 00:08:16.517 } 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@655 -- # es=1 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:16.517 15:56:14 app_cmdline -- app/cmdline.sh@1 -- # killprocess 58647 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@954 -- # '[' -z 58647 ']' 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@958 -- # kill -0 58647 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@959 -- # uname 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:16.517 15:56:14 app_cmdline -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58647 00:08:16.777 killing process with pid 58647 00:08:16.777 15:56:14 app_cmdline -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:16.777 15:56:14 app_cmdline -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:16.777 15:56:14 app_cmdline -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58647' 00:08:16.777 15:56:14 app_cmdline -- common/autotest_common.sh@973 -- # kill 58647 00:08:16.777 15:56:14 app_cmdline -- common/autotest_common.sh@978 -- # wait 58647 00:08:18.160 00:08:18.160 real 0m3.047s 00:08:18.160 user 0m3.378s 00:08:18.160 sys 0m0.424s 00:08:18.160 ************************************ 00:08:18.160 END TEST app_cmdline 00:08:18.160 ************************************ 00:08:18.160 15:56:16 app_cmdline -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:18.160 15:56:16 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:08:18.160 15:56:16 -- spdk/autotest.sh@177 -- # run_test version /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:08:18.160 15:56:16 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:18.160 15:56:16 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:18.160 15:56:16 -- common/autotest_common.sh@10 -- # set +x 00:08:18.160 ************************************ 00:08:18.160 START TEST version 00:08:18.160 ************************************ 00:08:18.160 15:56:16 version -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:08:18.160 * Looking for test storage... 00:08:18.160 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:08:18.160 15:56:16 version -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:08:18.160 15:56:16 version -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:08:18.160 15:56:16 version -- common/autotest_common.sh@1693 -- # lcov --version 00:08:18.428 15:56:16 version -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:08:18.428 15:56:16 version -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:18.428 15:56:16 version -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:18.428 15:56:16 version -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:18.428 15:56:16 version -- scripts/common.sh@336 -- # IFS=.-: 00:08:18.428 15:56:16 version -- scripts/common.sh@336 -- # read -ra ver1 00:08:18.428 15:56:16 version -- scripts/common.sh@337 -- # IFS=.-: 00:08:18.428 15:56:16 version -- scripts/common.sh@337 -- # read -ra ver2 00:08:18.428 15:56:16 version -- scripts/common.sh@338 -- # local 'op=<' 00:08:18.428 15:56:16 version -- scripts/common.sh@340 -- # ver1_l=2 00:08:18.428 15:56:16 version -- scripts/common.sh@341 -- # ver2_l=1 00:08:18.429 15:56:16 version -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:18.429 15:56:16 version -- scripts/common.sh@344 -- # case "$op" in 00:08:18.429 15:56:16 version -- scripts/common.sh@345 -- # : 1 00:08:18.429 15:56:16 version -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:18.429 15:56:16 version -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:18.429 15:56:16 version -- scripts/common.sh@365 -- # decimal 1 00:08:18.429 15:56:16 version -- scripts/common.sh@353 -- # local d=1 00:08:18.429 15:56:16 version -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:18.429 15:56:16 version -- scripts/common.sh@355 -- # echo 1 00:08:18.429 15:56:16 version -- scripts/common.sh@365 -- # ver1[v]=1 00:08:18.429 15:56:16 version -- scripts/common.sh@366 -- # decimal 2 00:08:18.429 15:56:16 version -- scripts/common.sh@353 -- # local d=2 00:08:18.429 15:56:16 version -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:18.429 15:56:16 version -- scripts/common.sh@355 -- # echo 2 00:08:18.429 15:56:16 version -- scripts/common.sh@366 -- # ver2[v]=2 00:08:18.429 15:56:16 version -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:18.429 15:56:16 version -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:18.429 15:56:16 version -- scripts/common.sh@368 -- # return 0 00:08:18.429 15:56:16 version -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:18.429 15:56:16 version -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:08:18.429 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:18.429 --rc genhtml_branch_coverage=1 00:08:18.429 --rc genhtml_function_coverage=1 00:08:18.429 --rc genhtml_legend=1 00:08:18.429 --rc geninfo_all_blocks=1 00:08:18.429 --rc geninfo_unexecuted_blocks=1 00:08:18.429 00:08:18.429 ' 00:08:18.429 15:56:16 version -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:08:18.429 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:18.429 --rc genhtml_branch_coverage=1 00:08:18.429 --rc genhtml_function_coverage=1 00:08:18.429 --rc genhtml_legend=1 00:08:18.429 --rc geninfo_all_blocks=1 00:08:18.429 --rc geninfo_unexecuted_blocks=1 00:08:18.429 00:08:18.429 ' 00:08:18.429 15:56:16 version -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:08:18.429 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:18.429 --rc genhtml_branch_coverage=1 00:08:18.429 --rc genhtml_function_coverage=1 00:08:18.429 --rc genhtml_legend=1 00:08:18.429 --rc geninfo_all_blocks=1 00:08:18.429 --rc geninfo_unexecuted_blocks=1 00:08:18.429 00:08:18.429 ' 00:08:18.429 15:56:16 version -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:08:18.429 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:18.429 --rc genhtml_branch_coverage=1 00:08:18.429 --rc genhtml_function_coverage=1 00:08:18.429 --rc genhtml_legend=1 00:08:18.429 --rc geninfo_all_blocks=1 00:08:18.429 --rc geninfo_unexecuted_blocks=1 00:08:18.429 00:08:18.429 ' 00:08:18.429 15:56:16 version -- app/version.sh@17 -- # get_header_version major 00:08:18.429 15:56:16 version -- app/version.sh@14 -- # cut -f2 00:08:18.429 15:56:16 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:08:18.429 15:56:16 version -- app/version.sh@14 -- # tr -d '"' 00:08:18.429 15:56:16 version -- app/version.sh@17 -- # major=25 00:08:18.429 15:56:16 version -- app/version.sh@18 -- # get_header_version minor 00:08:18.429 15:56:16 version -- app/version.sh@14 -- # cut -f2 00:08:18.429 15:56:16 version -- app/version.sh@14 -- # tr -d '"' 00:08:18.429 15:56:16 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:08:18.429 15:56:16 version -- app/version.sh@18 -- # minor=1 00:08:18.429 15:56:16 version -- app/version.sh@19 -- # get_header_version patch 00:08:18.429 15:56:16 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:08:18.429 15:56:16 version -- app/version.sh@14 -- # cut -f2 00:08:18.429 15:56:16 version -- app/version.sh@14 -- # tr -d '"' 00:08:18.429 15:56:16 version -- app/version.sh@19 -- # patch=0 00:08:18.429 15:56:16 version -- app/version.sh@20 -- # get_header_version suffix 00:08:18.429 15:56:16 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:08:18.429 15:56:16 version -- app/version.sh@14 -- # tr -d '"' 00:08:18.429 15:56:16 version -- app/version.sh@14 -- # cut -f2 00:08:18.429 15:56:16 version -- app/version.sh@20 -- # suffix=-pre 00:08:18.429 15:56:16 version -- app/version.sh@22 -- # version=25.1 00:08:18.429 15:56:16 version -- app/version.sh@25 -- # (( patch != 0 )) 00:08:18.429 15:56:16 version -- app/version.sh@28 -- # version=25.1rc0 00:08:18.429 15:56:16 version -- app/version.sh@30 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python 00:08:18.429 15:56:16 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:08:18.429 15:56:16 version -- app/version.sh@30 -- # py_version=25.1rc0 00:08:18.429 15:56:16 version -- app/version.sh@31 -- # [[ 25.1rc0 == \2\5\.\1\r\c\0 ]] 00:08:18.429 ************************************ 00:08:18.429 00:08:18.429 real 0m0.217s 00:08:18.429 user 0m0.146s 00:08:18.429 sys 0m0.099s 00:08:18.429 15:56:16 version -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:18.429 15:56:16 version -- common/autotest_common.sh@10 -- # set +x 00:08:18.429 END TEST version 00:08:18.429 ************************************ 00:08:18.429 15:56:16 -- spdk/autotest.sh@179 -- # '[' 0 -eq 1 ']' 00:08:18.429 15:56:16 -- spdk/autotest.sh@188 -- # [[ 1 -eq 1 ]] 00:08:18.429 15:56:16 -- spdk/autotest.sh@189 -- # run_test bdev_raid /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:08:18.429 15:56:16 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:18.429 15:56:16 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:18.429 15:56:16 -- common/autotest_common.sh@10 -- # set +x 00:08:18.429 ************************************ 00:08:18.429 START TEST bdev_raid 00:08:18.429 ************************************ 00:08:18.429 15:56:16 bdev_raid -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:08:18.429 * Looking for test storage... 00:08:18.429 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:08:18.429 15:56:16 bdev_raid -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:08:18.429 15:56:16 bdev_raid -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:08:18.429 15:56:16 bdev_raid -- common/autotest_common.sh@1693 -- # lcov --version 00:08:18.723 15:56:16 bdev_raid -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@336 -- # IFS=.-: 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@336 -- # read -ra ver1 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@337 -- # IFS=.-: 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@337 -- # read -ra ver2 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@338 -- # local 'op=<' 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@340 -- # ver1_l=2 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@341 -- # ver2_l=1 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@344 -- # case "$op" in 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@345 -- # : 1 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@365 -- # decimal 1 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@353 -- # local d=1 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@355 -- # echo 1 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@366 -- # decimal 2 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@353 -- # local d=2 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@355 -- # echo 2 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:18.723 15:56:16 bdev_raid -- scripts/common.sh@368 -- # return 0 00:08:18.723 15:56:16 bdev_raid -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:18.723 15:56:16 bdev_raid -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:08:18.723 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:18.723 --rc genhtml_branch_coverage=1 00:08:18.723 --rc genhtml_function_coverage=1 00:08:18.723 --rc genhtml_legend=1 00:08:18.723 --rc geninfo_all_blocks=1 00:08:18.723 --rc geninfo_unexecuted_blocks=1 00:08:18.723 00:08:18.723 ' 00:08:18.723 15:56:16 bdev_raid -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:08:18.723 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:18.723 --rc genhtml_branch_coverage=1 00:08:18.723 --rc genhtml_function_coverage=1 00:08:18.723 --rc genhtml_legend=1 00:08:18.723 --rc geninfo_all_blocks=1 00:08:18.723 --rc geninfo_unexecuted_blocks=1 00:08:18.723 00:08:18.723 ' 00:08:18.723 15:56:16 bdev_raid -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:08:18.723 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:18.723 --rc genhtml_branch_coverage=1 00:08:18.723 --rc genhtml_function_coverage=1 00:08:18.723 --rc genhtml_legend=1 00:08:18.723 --rc geninfo_all_blocks=1 00:08:18.723 --rc geninfo_unexecuted_blocks=1 00:08:18.723 00:08:18.723 ' 00:08:18.723 15:56:16 bdev_raid -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:08:18.723 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:18.723 --rc genhtml_branch_coverage=1 00:08:18.723 --rc genhtml_function_coverage=1 00:08:18.723 --rc genhtml_legend=1 00:08:18.723 --rc geninfo_all_blocks=1 00:08:18.723 --rc geninfo_unexecuted_blocks=1 00:08:18.723 00:08:18.723 ' 00:08:18.723 15:56:16 bdev_raid -- bdev/bdev_raid.sh@12 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:08:18.723 15:56:16 bdev_raid -- bdev/nbd_common.sh@6 -- # set -e 00:08:18.723 15:56:16 bdev_raid -- bdev/bdev_raid.sh@14 -- # rpc_py=rpc_cmd 00:08:18.723 15:56:16 bdev_raid -- bdev/bdev_raid.sh@946 -- # mkdir -p /raidtest 00:08:18.723 15:56:16 bdev_raid -- bdev/bdev_raid.sh@947 -- # trap 'cleanup; exit 1' EXIT 00:08:18.723 15:56:16 bdev_raid -- bdev/bdev_raid.sh@949 -- # base_blocklen=512 00:08:18.723 15:56:16 bdev_raid -- bdev/bdev_raid.sh@951 -- # run_test raid1_resize_data_offset_test raid_resize_data_offset_test 00:08:18.723 15:56:16 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:18.723 15:56:16 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:18.723 15:56:16 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:18.723 ************************************ 00:08:18.723 START TEST raid1_resize_data_offset_test 00:08:18.723 ************************************ 00:08:18.723 15:56:16 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1129 -- # raid_resize_data_offset_test 00:08:18.723 15:56:16 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@917 -- # raid_pid=58819 00:08:18.723 15:56:16 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@918 -- # echo 'Process raid pid: 58819' 00:08:18.723 Process raid pid: 58819 00:08:18.723 15:56:16 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@916 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:18.723 15:56:16 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@919 -- # waitforlisten 58819 00:08:18.723 15:56:16 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@835 -- # '[' -z 58819 ']' 00:08:18.723 15:56:16 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:18.723 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:18.723 15:56:16 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:18.724 15:56:16 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:18.724 15:56:16 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:18.724 15:56:16 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.724 [2024-11-20 15:56:16.821014] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:18.724 [2024-11-20 15:56:16.821706] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:18.984 [2024-11-20 15:56:16.983700] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:18.984 [2024-11-20 15:56:17.087237] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:18.984 [2024-11-20 15:56:17.225777] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:18.984 [2024-11-20 15:56:17.225970] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:19.553 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:19.553 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@868 -- # return 0 00:08:19.553 15:56:17 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@922 -- # rpc_cmd bdev_malloc_create -b malloc0 64 512 -o 16 00:08:19.553 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.553 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.553 malloc0 00:08:19.553 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.553 15:56:17 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@923 -- # rpc_cmd bdev_malloc_create -b malloc1 64 512 -o 16 00:08:19.553 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.553 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.813 malloc1 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@924 -- # rpc_cmd bdev_null_create null0 64 512 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.813 null0 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@926 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''malloc0 malloc1 null0'\''' -s 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.813 [2024-11-20 15:56:17.820759] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc0 is claimed 00:08:19.813 [2024-11-20 15:56:17.822697] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:19.813 [2024-11-20 15:56:17.822747] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev null0 is claimed 00:08:19.813 [2024-11-20 15:56:17.822870] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:08:19.813 [2024-11-20 15:56:17.822883] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 129024, blocklen 512 00:08:19.813 [2024-11-20 15:56:17.823137] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:08:19.813 [2024-11-20 15:56:17.823269] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:08:19.813 [2024-11-20 15:56:17.823279] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:08:19.813 [2024-11-20 15:56:17.823411] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # (( 2048 == 2048 )) 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@931 -- # rpc_cmd bdev_null_delete null0 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.813 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.814 [2024-11-20 15:56:17.872790] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: null0 00:08:19.814 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.814 15:56:17 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@935 -- # rpc_cmd bdev_malloc_create -b malloc2 512 512 -o 30 00:08:19.814 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.814 15:56:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.074 malloc2 00:08:20.074 15:56:18 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.074 15:56:18 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@936 -- # rpc_cmd bdev_raid_add_base_bdev Raid malloc2 00:08:20.074 15:56:18 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.074 15:56:18 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.074 [2024-11-20 15:56:18.248869] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:20.074 [2024-11-20 15:56:18.260485] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:20.074 15:56:18 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.074 [2024-11-20 15:56:18.262346] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev Raid 00:08:20.074 15:56:18 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:08:20.074 15:56:18 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.074 15:56:18 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.074 15:56:18 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.074 15:56:18 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.074 15:56:18 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # (( 2070 == 2070 )) 00:08:20.074 15:56:18 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@941 -- # killprocess 58819 00:08:20.074 15:56:18 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@954 -- # '[' -z 58819 ']' 00:08:20.074 15:56:18 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@958 -- # kill -0 58819 00:08:20.074 15:56:18 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@959 -- # uname 00:08:20.074 15:56:18 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:20.074 15:56:18 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58819 00:08:20.334 killing process with pid 58819 00:08:20.334 15:56:18 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:20.334 15:56:18 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:20.334 15:56:18 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58819' 00:08:20.334 15:56:18 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@973 -- # kill 58819 00:08:20.334 15:56:18 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@978 -- # wait 58819 00:08:20.334 [2024-11-20 15:56:18.323043] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:20.334 [2024-11-20 15:56:18.323329] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev Raid: Operation canceled 00:08:20.334 [2024-11-20 15:56:18.323528] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:20.334 [2024-11-20 15:56:18.323549] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: malloc2 00:08:20.334 [2024-11-20 15:56:18.346040] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:20.334 [2024-11-20 15:56:18.346428] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:20.334 [2024-11-20 15:56:18.346450] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:08:21.275 [2024-11-20 15:56:19.438372] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:22.218 15:56:20 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@943 -- # return 0 00:08:22.218 00:08:22.218 real 0m3.393s 00:08:22.218 user 0m3.414s 00:08:22.218 sys 0m0.386s 00:08:22.218 ************************************ 00:08:22.218 END TEST raid1_resize_data_offset_test 00:08:22.218 ************************************ 00:08:22.218 15:56:20 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:22.218 15:56:20 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.218 15:56:20 bdev_raid -- bdev/bdev_raid.sh@953 -- # run_test raid0_resize_superblock_test raid_resize_superblock_test 0 00:08:22.218 15:56:20 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:08:22.218 15:56:20 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:22.218 15:56:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:22.218 ************************************ 00:08:22.218 START TEST raid0_resize_superblock_test 00:08:22.218 ************************************ 00:08:22.218 15:56:20 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1129 -- # raid_resize_superblock_test 0 00:08:22.218 15:56:20 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=0 00:08:22.218 Process raid pid: 58896 00:08:22.218 15:56:20 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=58896 00:08:22.218 15:56:20 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 58896' 00:08:22.218 15:56:20 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 58896 00:08:22.218 15:56:20 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 58896 ']' 00:08:22.218 15:56:20 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:22.218 15:56:20 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:22.218 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:22.218 15:56:20 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:22.218 15:56:20 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:22.218 15:56:20 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.218 15:56:20 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:22.218 [2024-11-20 15:56:20.276910] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:22.218 [2024-11-20 15:56:20.277031] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:22.218 [2024-11-20 15:56:20.437504] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:22.478 [2024-11-20 15:56:20.539965] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:22.478 [2024-11-20 15:56:20.678149] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:22.478 [2024-11-20 15:56:20.678188] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:23.043 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:23.043 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:08:23.043 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:08:23.043 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.043 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.300 malloc0 00:08:23.300 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.300 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:08:23.300 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.300 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.300 [2024-11-20 15:56:21.491634] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:08:23.300 [2024-11-20 15:56:21.491713] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:23.300 [2024-11-20 15:56:21.491734] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:23.300 [2024-11-20 15:56:21.491746] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:23.300 [2024-11-20 15:56:21.493892] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:23.300 [2024-11-20 15:56:21.493930] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:08:23.300 pt0 00:08:23.300 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.300 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:08:23.300 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.300 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.558 8f20cd72-0b71-47b2-83b3-c16b63622236 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.558 b597e4f1-6228-49e2-879a-b1760e71731c 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.558 3ed1e2bf-48c0-4ecd-be9f-ee8c5cd2fc87 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@870 -- # rpc_cmd bdev_raid_create -n Raid -r 0 -z 64 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.558 [2024-11-20 15:56:21.580482] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev b597e4f1-6228-49e2-879a-b1760e71731c is claimed 00:08:23.558 [2024-11-20 15:56:21.580563] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 3ed1e2bf-48c0-4ecd-be9f-ee8c5cd2fc87 is claimed 00:08:23.558 [2024-11-20 15:56:21.580702] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:08:23.558 [2024-11-20 15:56:21.580717] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 245760, blocklen 512 00:08:23.558 [2024-11-20 15:56:21.580963] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:23.558 [2024-11-20 15:56:21.581137] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:08:23.558 [2024-11-20 15:56:21.581147] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:08:23.558 [2024-11-20 15:56:21.581284] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # jq '.[].num_blocks' 00:08:23.558 [2024-11-20 15:56:21.656752] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # (( 245760 == 245760 )) 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.558 [2024-11-20 15:56:21.692701] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:08:23.558 [2024-11-20 15:56:21.692730] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'b597e4f1-6228-49e2-879a-b1760e71731c' was resized: old size 131072, new size 204800 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.558 [2024-11-20 15:56:21.700613] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:08:23.558 [2024-11-20 15:56:21.700638] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '3ed1e2bf-48c0-4ecd-be9f-ee8c5cd2fc87' was resized: old size 131072, new size 204800 00:08:23.558 [2024-11-20 15:56:21.700660] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 245760 to 393216 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:08:23.558 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # jq '.[].num_blocks' 00:08:23.559 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.559 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.559 [2024-11-20 15:56:21.776769] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:23.559 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.559 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:08:23.559 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:08:23.559 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # (( 393216 == 393216 )) 00:08:23.559 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:08:23.559 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.559 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.816 [2024-11-20 15:56:21.808538] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:08:23.816 [2024-11-20 15:56:21.808605] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:08:23.816 [2024-11-20 15:56:21.808618] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:23.816 [2024-11-20 15:56:21.808631] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:08:23.816 [2024-11-20 15:56:21.808741] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:23.816 [2024-11-20 15:56:21.808775] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:23.816 [2024-11-20 15:56:21.808785] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:08:23.816 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.816 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:08:23.816 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.816 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.816 [2024-11-20 15:56:21.816483] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:08:23.816 [2024-11-20 15:56:21.816528] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:23.816 [2024-11-20 15:56:21.816545] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008180 00:08:23.816 [2024-11-20 15:56:21.816555] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:23.816 [2024-11-20 15:56:21.818688] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:23.816 [2024-11-20 15:56:21.818720] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:08:23.817 [2024-11-20 15:56:21.820245] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev b597e4f1-6228-49e2-879a-b1760e71731c 00:08:23.817 [2024-11-20 15:56:21.820306] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev b597e4f1-6228-49e2-879a-b1760e71731c is claimed 00:08:23.817 [2024-11-20 15:56:21.820402] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 3ed1e2bf-48c0-4ecd-be9f-ee8c5cd2fc87 00:08:23.817 [2024-11-20 15:56:21.820419] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 3ed1e2bf-48c0-4ecd-be9f-ee8c5cd2fc87 is claimed 00:08:23.817 [2024-11-20 15:56:21.820530] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev 3ed1e2bf-48c0-4ecd-be9f-ee8c5cd2fc87 (2) smaller than existing raid bdev Raid (3) 00:08:23.817 [2024-11-20 15:56:21.820552] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev b597e4f1-6228-49e2-879a-b1760e71731c: File exists 00:08:23.817 [2024-11-20 15:56:21.820592] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:08:23.817 [2024-11-20 15:56:21.820603] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 393216, blocklen 512 00:08:23.817 [2024-11-20 15:56:21.820847] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:08:23.817 [2024-11-20 15:56:21.820989] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:08:23.817 [2024-11-20 15:56:21.821001] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007b00 00:08:23.817 [2024-11-20 15:56:21.821183] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:23.817 pt0 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # jq '.[].num_blocks' 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.817 [2024-11-20 15:56:21.836851] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # (( 393216 == 393216 )) 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 58896 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 58896 ']' 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@958 -- # kill -0 58896 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@959 -- # uname 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58896 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:23.817 killing process with pid 58896 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58896' 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@973 -- # kill 58896 00:08:23.817 [2024-11-20 15:56:21.898594] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:23.817 [2024-11-20 15:56:21.898660] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:23.817 15:56:21 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@978 -- # wait 58896 00:08:23.817 [2024-11-20 15:56:21.898713] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:23.817 [2024-11-20 15:56:21.898722] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Raid, state offline 00:08:24.806 [2024-11-20 15:56:22.790181] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:25.370 15:56:23 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:08:25.370 00:08:25.370 real 0m3.293s 00:08:25.370 user 0m3.486s 00:08:25.370 sys 0m0.413s 00:08:25.370 15:56:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:25.370 ************************************ 00:08:25.370 15:56:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.370 END TEST raid0_resize_superblock_test 00:08:25.370 ************************************ 00:08:25.370 15:56:23 bdev_raid -- bdev/bdev_raid.sh@954 -- # run_test raid1_resize_superblock_test raid_resize_superblock_test 1 00:08:25.371 15:56:23 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:08:25.371 15:56:23 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:25.371 15:56:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:25.371 ************************************ 00:08:25.371 START TEST raid1_resize_superblock_test 00:08:25.371 ************************************ 00:08:25.371 15:56:23 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1129 -- # raid_resize_superblock_test 1 00:08:25.371 15:56:23 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=1 00:08:25.371 15:56:23 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=58978 00:08:25.371 Process raid pid: 58978 00:08:25.371 15:56:23 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:25.371 15:56:23 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 58978' 00:08:25.371 15:56:23 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 58978 00:08:25.371 15:56:23 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 58978 ']' 00:08:25.371 15:56:23 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:25.371 15:56:23 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:25.371 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:25.371 15:56:23 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:25.371 15:56:23 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:25.371 15:56:23 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.371 [2024-11-20 15:56:23.607794] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:25.371 [2024-11-20 15:56:23.607913] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:25.628 [2024-11-20 15:56:23.763850] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:25.628 [2024-11-20 15:56:23.845584] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:25.885 [2024-11-20 15:56:23.954695] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:25.885 [2024-11-20 15:56:23.954725] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:26.449 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:26.450 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:08:26.450 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:08:26.450 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.450 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.708 malloc0 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.708 [2024-11-20 15:56:24.807889] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:08:26.708 [2024-11-20 15:56:24.807947] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:26.708 [2024-11-20 15:56:24.807965] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:26.708 [2024-11-20 15:56:24.807976] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:26.708 [2024-11-20 15:56:24.809764] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:26.708 [2024-11-20 15:56:24.809799] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:08:26.708 pt0 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.708 8062a919-0224-431e-a84c-088e1de2e638 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.708 95f68c42-3409-4c24-aba5-ce0a305fdad4 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.708 cd99848c-e644-4334-945f-71eaf87c036f 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@871 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.708 [2024-11-20 15:56:24.890539] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 95f68c42-3409-4c24-aba5-ce0a305fdad4 is claimed 00:08:26.708 [2024-11-20 15:56:24.890623] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev cd99848c-e644-4334-945f-71eaf87c036f is claimed 00:08:26.708 [2024-11-20 15:56:24.890753] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:08:26.708 [2024-11-20 15:56:24.890766] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 122880, blocklen 512 00:08:26.708 [2024-11-20 15:56:24.890983] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:26.708 [2024-11-20 15:56:24.891127] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:08:26.708 [2024-11-20 15:56:24.891134] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:08:26.708 [2024-11-20 15:56:24.891255] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.708 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # jq '.[].num_blocks' 00:08:26.966 [2024-11-20 15:56:24.962769] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # (( 122880 == 122880 )) 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.966 [2024-11-20 15:56:24.986722] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:08:26.966 [2024-11-20 15:56:24.986749] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '95f68c42-3409-4c24-aba5-ce0a305fdad4' was resized: old size 131072, new size 204800 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.966 [2024-11-20 15:56:24.994687] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:08:26.966 [2024-11-20 15:56:24.994712] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'cd99848c-e644-4334-945f-71eaf87c036f' was resized: old size 131072, new size 204800 00:08:26.966 [2024-11-20 15:56:24.994732] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 122880 to 196608 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.966 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:08:26.966 15:56:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:08:26.966 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.966 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.966 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.966 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:08:26.966 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:08:26.966 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:08:26.966 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.966 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.966 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.966 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # jq '.[].num_blocks' 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.967 [2024-11-20 15:56:25.066773] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # (( 196608 == 196608 )) 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.967 [2024-11-20 15:56:25.090588] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:08:26.967 [2024-11-20 15:56:25.090654] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:08:26.967 [2024-11-20 15:56:25.090686] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:08:26.967 [2024-11-20 15:56:25.090813] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:26.967 [2024-11-20 15:56:25.090959] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:26.967 [2024-11-20 15:56:25.091010] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:26.967 [2024-11-20 15:56:25.091020] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.967 [2024-11-20 15:56:25.098525] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:08:26.967 [2024-11-20 15:56:25.098570] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:26.967 [2024-11-20 15:56:25.098587] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008180 00:08:26.967 [2024-11-20 15:56:25.098599] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:26.967 [2024-11-20 15:56:25.100417] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:26.967 [2024-11-20 15:56:25.100449] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:08:26.967 [2024-11-20 15:56:25.101752] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 95f68c42-3409-4c24-aba5-ce0a305fdad4 00:08:26.967 [2024-11-20 15:56:25.101809] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 95f68c42-3409-4c24-aba5-ce0a305fdad4 is claimed 00:08:26.967 [2024-11-20 15:56:25.101889] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev cd99848c-e644-4334-945f-71eaf87c036f 00:08:26.967 [2024-11-20 15:56:25.101903] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev cd99848c-e644-4334-945f-71eaf87c036f is claimed 00:08:26.967 [2024-11-20 15:56:25.101994] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev cd99848c-e644-4334-945f-71eaf87c036f (2) smaller than existing raid bdev Raid (3) 00:08:26.967 [2024-11-20 15:56:25.102018] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev 95f68c42-3409-4c24-aba5-ce0a305fdad4: File exists 00:08:26.967 [2024-11-20 15:56:25.102051] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:08:26.967 [2024-11-20 15:56:25.102060] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:26.967 [2024-11-20 15:56:25.102250] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:08:26.967 [2024-11-20 15:56:25.102371] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:08:26.967 [2024-11-20 15:56:25.102381] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007b00 00:08:26.967 [2024-11-20 15:56:25.102530] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:26.967 pt0 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # jq '.[].num_blocks' 00:08:26.967 [2024-11-20 15:56:25.118768] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # (( 196608 == 196608 )) 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 58978 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 58978 ']' 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@958 -- # kill -0 58978 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@959 -- # uname 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58978 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58978' 00:08:26.967 killing process with pid 58978 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@973 -- # kill 58978 00:08:26.967 [2024-11-20 15:56:25.169157] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:26.967 [2024-11-20 15:56:25.169227] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:26.967 15:56:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@978 -- # wait 58978 00:08:26.967 [2024-11-20 15:56:25.169276] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:26.967 [2024-11-20 15:56:25.169284] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Raid, state offline 00:08:27.899 [2024-11-20 15:56:25.859975] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:28.242 15:56:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:08:28.242 00:08:28.242 real 0m2.900s 00:08:28.242 user 0m3.171s 00:08:28.242 sys 0m0.363s 00:08:28.242 ************************************ 00:08:28.242 END TEST raid1_resize_superblock_test 00:08:28.242 ************************************ 00:08:28.242 15:56:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:28.242 15:56:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.242 15:56:26 bdev_raid -- bdev/bdev_raid.sh@956 -- # uname -s 00:08:28.242 15:56:26 bdev_raid -- bdev/bdev_raid.sh@956 -- # '[' Linux = Linux ']' 00:08:28.242 15:56:26 bdev_raid -- bdev/bdev_raid.sh@956 -- # modprobe -n nbd 00:08:28.242 15:56:26 bdev_raid -- bdev/bdev_raid.sh@957 -- # has_nbd=true 00:08:28.242 15:56:26 bdev_raid -- bdev/bdev_raid.sh@958 -- # modprobe nbd 00:08:28.499 15:56:26 bdev_raid -- bdev/bdev_raid.sh@959 -- # run_test raid_function_test_raid0 raid_function_test raid0 00:08:28.500 15:56:26 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:08:28.500 15:56:26 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:28.500 15:56:26 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:28.500 ************************************ 00:08:28.500 START TEST raid_function_test_raid0 00:08:28.500 ************************************ 00:08:28.500 15:56:26 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1129 -- # raid_function_test raid0 00:08:28.500 15:56:26 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@64 -- # local raid_level=raid0 00:08:28.500 15:56:26 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:08:28.500 15:56:26 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:08:28.500 15:56:26 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@69 -- # raid_pid=59064 00:08:28.500 Process raid pid: 59064 00:08:28.500 15:56:26 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 59064' 00:08:28.500 15:56:26 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@71 -- # waitforlisten 59064 00:08:28.500 15:56:26 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@835 -- # '[' -z 59064 ']' 00:08:28.500 15:56:26 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:28.500 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:28.500 15:56:26 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:28.500 15:56:26 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:28.500 15:56:26 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:28.500 15:56:26 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:28.500 15:56:26 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:08:28.500 [2024-11-20 15:56:26.560458] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:28.500 [2024-11-20 15:56:26.560581] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:28.500 [2024-11-20 15:56:26.716714] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:28.757 [2024-11-20 15:56:26.801180] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:28.757 [2024-11-20 15:56:26.912106] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:28.757 [2024-11-20 15:56:26.912142] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@868 -- # return 0 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:08:29.321 Base_1 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:08:29.321 Base_2 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''Base_1 Base_2'\''' -n raid 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:08:29.321 [2024-11-20 15:56:27.464989] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:08:29.321 [2024-11-20 15:56:27.466543] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:08:29.321 [2024-11-20 15:56:27.466609] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:08:29.321 [2024-11-20 15:56:27.466620] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:08:29.321 [2024-11-20 15:56:27.466864] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:29.321 [2024-11-20 15:56:27.466976] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:08:29.321 [2024-11-20 15:56:27.466983] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000007780 00:08:29.321 [2024-11-20 15:56:27.467107] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # local bdev_list 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # local nbd_list 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@12 -- # local i 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:08:29.321 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:08:29.579 [2024-11-20 15:56:27.693057] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:08:29.579 /dev/nbd0 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@873 -- # local i 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@877 -- # break 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:08:29.579 1+0 records in 00:08:29.579 1+0 records out 00:08:29.579 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000237355 s, 17.3 MB/s 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@890 -- # size=4096 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@893 -- # return 0 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:08:29.579 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:08:29.837 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:08:29.837 { 00:08:29.837 "nbd_device": "/dev/nbd0", 00:08:29.837 "bdev_name": "raid" 00:08:29.837 } 00:08:29.837 ]' 00:08:29.837 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[ 00:08:29.837 { 00:08:29.837 "nbd_device": "/dev/nbd0", 00:08:29.837 "bdev_name": "raid" 00:08:29.837 } 00:08:29.837 ]' 00:08:29.837 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:29.837 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:08:29.837 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:08:29.837 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:29.837 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=1 00:08:29.837 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 1 00:08:29.837 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # count=1 00:08:29.837 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:08:29.837 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:08:29.837 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:08:29.837 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:08:29.837 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@19 -- # local blksize 00:08:29.837 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:08:29.837 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:08:29.837 15:56:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:08:29.837 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # blksize=512 00:08:29.837 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:08:29.837 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:08:29.837 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:08:29.837 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:08:29.837 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:08:29.837 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:08:29.837 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:08:29.837 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:08:29.837 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:08:29.837 4096+0 records in 00:08:29.838 4096+0 records out 00:08:29.838 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0205246 s, 102 MB/s 00:08:29.838 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:08:30.095 4096+0 records in 00:08:30.095 4096+0 records out 00:08:30.095 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.232892 s, 9.0 MB/s 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:08:30.095 128+0 records in 00:08:30.095 128+0 records out 00:08:30.095 65536 bytes (66 kB, 64 KiB) copied, 0.000310557 s, 211 MB/s 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:08:30.095 2035+0 records in 00:08:30.095 2035+0 records out 00:08:30.095 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.00719606 s, 145 MB/s 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:08:30.095 456+0 records in 00:08:30.095 456+0 records out 00:08:30.095 233472 bytes (233 kB, 228 KiB) copied, 0.00184746 s, 126 MB/s 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@52 -- # return 0 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # local nbd_list 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@51 -- # local i 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:08:30.095 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:08:30.353 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:08:30.353 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:08:30.353 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:08:30.353 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:08:30.353 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:08:30.353 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:08:30.353 [2024-11-20 15:56:28.538698] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:30.353 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@41 -- # break 00:08:30.353 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@45 -- # return 0 00:08:30.353 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:08:30.353 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:08:30.353 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[]' 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo '' 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # true 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=0 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 0 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # count=0 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@97 -- # killprocess 59064 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@954 -- # '[' -z 59064 ']' 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@958 -- # kill -0 59064 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@959 -- # uname 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 59064 00:08:30.610 killing process with pid 59064 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 59064' 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@973 -- # kill 59064 00:08:30.610 15:56:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@978 -- # wait 59064 00:08:30.610 [2024-11-20 15:56:28.816049] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:30.610 [2024-11-20 15:56:28.816143] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:30.610 [2024-11-20 15:56:28.816195] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:30.610 [2024-11-20 15:56:28.816211] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid, state offline 00:08:30.867 [2024-11-20 15:56:28.950083] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:31.815 ************************************ 00:08:31.815 END TEST raid_function_test_raid0 00:08:31.815 ************************************ 00:08:31.815 15:56:29 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@99 -- # return 0 00:08:31.815 00:08:31.815 real 0m3.211s 00:08:31.815 user 0m3.867s 00:08:31.815 sys 0m0.712s 00:08:31.815 15:56:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:31.815 15:56:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:08:31.815 15:56:29 bdev_raid -- bdev/bdev_raid.sh@960 -- # run_test raid_function_test_concat raid_function_test concat 00:08:31.815 15:56:29 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:08:31.815 15:56:29 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:31.815 15:56:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:31.815 ************************************ 00:08:31.815 START TEST raid_function_test_concat 00:08:31.815 ************************************ 00:08:31.815 15:56:29 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1129 -- # raid_function_test concat 00:08:31.815 15:56:29 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@64 -- # local raid_level=concat 00:08:31.815 15:56:29 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:08:31.815 15:56:29 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:08:31.815 Process raid pid: 59182 00:08:31.815 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:31.815 15:56:29 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@69 -- # raid_pid=59182 00:08:31.815 15:56:29 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 59182' 00:08:31.815 15:56:29 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@71 -- # waitforlisten 59182 00:08:31.815 15:56:29 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@835 -- # '[' -z 59182 ']' 00:08:31.815 15:56:29 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:31.815 15:56:29 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:31.815 15:56:29 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:31.815 15:56:29 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:31.815 15:56:29 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:31.815 15:56:29 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:08:31.815 [2024-11-20 15:56:29.808130] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:31.815 [2024-11-20 15:56:29.808248] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:31.815 [2024-11-20 15:56:29.966624] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:32.074 [2024-11-20 15:56:30.065258] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:32.074 [2024-11-20 15:56:30.201327] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:32.074 [2024-11-20 15:56:30.201366] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@868 -- # return 0 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:08:32.638 Base_1 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:08:32.638 Base_2 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''Base_1 Base_2'\''' -n raid 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:08:32.638 [2024-11-20 15:56:30.736420] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:08:32.638 [2024-11-20 15:56:30.738350] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:08:32.638 [2024-11-20 15:56:30.738507] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:08:32.638 [2024-11-20 15:56:30.738524] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:08:32.638 [2024-11-20 15:56:30.738797] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:32.638 [2024-11-20 15:56:30.738926] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:08:32.638 [2024-11-20 15:56:30.738935] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000007780 00:08:32.638 [2024-11-20 15:56:30.739069] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@12 -- # local i 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:08:32.638 15:56:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:08:32.896 [2024-11-20 15:56:31.012519] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:08:32.896 /dev/nbd0 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@873 -- # local i 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@877 -- # break 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:08:32.896 1+0 records in 00:08:32.896 1+0 records out 00:08:32.896 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000379262 s, 10.8 MB/s 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@890 -- # size=4096 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@893 -- # return 0 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:08:32.896 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:08:33.154 { 00:08:33.154 "nbd_device": "/dev/nbd0", 00:08:33.154 "bdev_name": "raid" 00:08:33.154 } 00:08:33.154 ]' 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[ 00:08:33.154 { 00:08:33.154 "nbd_device": "/dev/nbd0", 00:08:33.154 "bdev_name": "raid" 00:08:33.154 } 00:08:33.154 ]' 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=1 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 1 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # count=1 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@19 -- # local blksize 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # blksize=512 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:08:33.154 4096+0 records in 00:08:33.154 4096+0 records out 00:08:33.154 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.022169 s, 94.6 MB/s 00:08:33.154 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:08:33.411 4096+0 records in 00:08:33.411 4096+0 records out 00:08:33.411 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.232038 s, 9.0 MB/s 00:08:33.411 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:08:33.411 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:08:33.411 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:08:33.411 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:08:33.411 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:08:33.411 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:08:33.411 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:08:33.412 128+0 records in 00:08:33.412 128+0 records out 00:08:33.412 65536 bytes (66 kB, 64 KiB) copied, 0.000829757 s, 79.0 MB/s 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:08:33.412 2035+0 records in 00:08:33.412 2035+0 records out 00:08:33.412 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.00562974 s, 185 MB/s 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:08:33.412 456+0 records in 00:08:33.412 456+0 records out 00:08:33.412 233472 bytes (233 kB, 228 KiB) copied, 0.00194831 s, 120 MB/s 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@52 -- # return 0 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@51 -- # local i 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:08:33.412 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:08:33.670 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:08:33.670 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:08:33.670 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:08:33.670 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:08:33.670 [2024-11-20 15:56:31.836351] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:33.670 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:08:33.670 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:08:33.670 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@41 -- # break 00:08:33.670 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@45 -- # return 0 00:08:33.670 15:56:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:08:33.670 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:08:33.670 15:56:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo '' 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # true 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=0 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 0 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # count=0 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@97 -- # killprocess 59182 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@954 -- # '[' -z 59182 ']' 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@958 -- # kill -0 59182 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@959 -- # uname 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 59182 00:08:33.927 killing process with pid 59182 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@972 -- # echo 'killing process with pid 59182' 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@973 -- # kill 59182 00:08:33.927 [2024-11-20 15:56:32.112530] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:33.927 15:56:32 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@978 -- # wait 59182 00:08:33.927 [2024-11-20 15:56:32.112619] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:33.927 [2024-11-20 15:56:32.112680] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:33.927 [2024-11-20 15:56:32.112698] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid, state offline 00:08:34.184 [2024-11-20 15:56:32.241555] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:34.746 15:56:32 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@99 -- # return 0 00:08:34.746 00:08:34.746 real 0m3.217s 00:08:34.746 user 0m3.917s 00:08:34.746 sys 0m0.700s 00:08:34.746 15:56:32 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:34.746 ************************************ 00:08:34.746 END TEST raid_function_test_concat 00:08:34.746 ************************************ 00:08:34.746 15:56:32 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:08:35.003 15:56:32 bdev_raid -- bdev/bdev_raid.sh@963 -- # run_test raid0_resize_test raid_resize_test 0 00:08:35.003 15:56:32 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:08:35.003 15:56:32 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:35.003 15:56:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:35.003 ************************************ 00:08:35.003 START TEST raid0_resize_test 00:08:35.003 ************************************ 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1129 -- # raid_resize_test 0 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=0 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:08:35.003 Process raid pid: 59298 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=59298 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 59298' 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 59298 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@835 -- # '[' -z 59298 ']' 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:35.003 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:35.003 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.003 [2024-11-20 15:56:33.064299] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:35.003 [2024-11-20 15:56:33.064571] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:35.003 [2024-11-20 15:56:33.225251] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:35.260 [2024-11-20 15:56:33.325517] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:35.260 [2024-11-20 15:56:33.462171] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:35.260 [2024-11-20 15:56:33.462205] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:35.824 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:35.824 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@868 -- # return 0 00:08:35.824 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:08:35.824 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.824 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.824 Base_1 00:08:35.824 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.824 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:08:35.824 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.824 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.824 Base_2 00:08:35.824 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.825 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 0 -eq 0 ']' 00:08:35.825 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@350 -- # rpc_cmd bdev_raid_create -z 64 -r 0 -b ''\''Base_1 Base_2'\''' -n Raid 00:08:35.825 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.825 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.825 [2024-11-20 15:56:33.972131] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:08:35.825 [2024-11-20 15:56:33.974039] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:08:35.825 [2024-11-20 15:56:33.974182] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:08:35.825 [2024-11-20 15:56:33.974258] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:08:35.825 [2024-11-20 15:56:33.974537] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:08:35.825 [2024-11-20 15:56:33.974730] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:08:35.825 [2024-11-20 15:56:33.974743] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:08:35.825 [2024-11-20 15:56:33.974885] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:35.825 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.825 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:08:35.825 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.825 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.825 [2024-11-20 15:56:33.980126] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:08:35.825 [2024-11-20 15:56:33.980233] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:08:35.825 true 00:08:35.825 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.825 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:35.825 15:56:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:08:35.825 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.825 15:56:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.825 [2024-11-20 15:56:33.992311] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=131072 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=64 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 0 -eq 0 ']' 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@362 -- # expected_size=64 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 64 '!=' 64 ']' 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.825 [2024-11-20 15:56:34.024115] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:08:35.825 [2024-11-20 15:56:34.024220] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:08:35.825 [2024-11-20 15:56:34.024312] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 131072 to 262144 00:08:35.825 true 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.825 [2024-11-20 15:56:34.036300] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=262144 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=128 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 0 -eq 0 ']' 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@378 -- # expected_size=128 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 128 '!=' 128 ']' 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 59298 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@954 -- # '[' -z 59298 ']' 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@958 -- # kill -0 59298 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@959 -- # uname 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:35.825 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 59298 00:08:36.082 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:36.082 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:36.082 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 59298' 00:08:36.082 killing process with pid 59298 00:08:36.082 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@973 -- # kill 59298 00:08:36.082 [2024-11-20 15:56:34.091591] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:36.082 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@978 -- # wait 59298 00:08:36.082 [2024-11-20 15:56:34.091789] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:36.082 [2024-11-20 15:56:34.092183] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:36.082 [2024-11-20 15:56:34.092301] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:08:36.082 [2024-11-20 15:56:34.103406] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:36.646 ************************************ 00:08:36.646 END TEST raid0_resize_test 00:08:36.646 ************************************ 00:08:36.646 15:56:34 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:08:36.646 00:08:36.646 real 0m1.815s 00:08:36.647 user 0m1.998s 00:08:36.647 sys 0m0.259s 00:08:36.647 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:36.647 15:56:34 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.647 15:56:34 bdev_raid -- bdev/bdev_raid.sh@964 -- # run_test raid1_resize_test raid_resize_test 1 00:08:36.647 15:56:34 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:08:36.647 15:56:34 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:36.647 15:56:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:36.647 ************************************ 00:08:36.647 START TEST raid1_resize_test 00:08:36.647 ************************************ 00:08:36.647 Process raid pid: 59354 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1129 -- # raid_resize_test 1 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=1 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=59354 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 59354' 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 59354 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@835 -- # '[' -z 59354 ']' 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:36.647 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:36.647 15:56:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.904 [2024-11-20 15:56:34.913975] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:36.904 [2024-11-20 15:56:34.914087] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:36.904 [2024-11-20 15:56:35.074979] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:37.162 [2024-11-20 15:56:35.173694] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:37.162 [2024-11-20 15:56:35.309738] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:37.162 [2024-11-20 15:56:35.309787] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@868 -- # return 0 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.727 Base_1 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.727 Base_2 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 1 -eq 0 ']' 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@352 -- # rpc_cmd bdev_raid_create -r 1 -b ''\''Base_1 Base_2'\''' -n Raid 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.727 [2024-11-20 15:56:35.797431] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:08:37.727 [2024-11-20 15:56:35.799323] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:08:37.727 [2024-11-20 15:56:35.799378] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:08:37.727 [2024-11-20 15:56:35.799389] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:08:37.727 [2024-11-20 15:56:35.799638] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:08:37.727 [2024-11-20 15:56:35.799769] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:08:37.727 [2024-11-20 15:56:35.799779] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:08:37.727 [2024-11-20 15:56:35.799912] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.727 [2024-11-20 15:56:35.805424] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:08:37.727 [2024-11-20 15:56:35.805449] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:08:37.727 true 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:08:37.727 [2024-11-20 15:56:35.817599] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=65536 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=32 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 1 -eq 0 ']' 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@364 -- # expected_size=32 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 32 '!=' 32 ']' 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.727 [2024-11-20 15:56:35.849422] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:08:37.727 [2024-11-20 15:56:35.849533] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:08:37.727 [2024-11-20 15:56:35.849618] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 65536 to 131072 00:08:37.727 true 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.727 [2024-11-20 15:56:35.861599] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=131072 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=64 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 1 -eq 0 ']' 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@380 -- # expected_size=64 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 64 '!=' 64 ']' 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 59354 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@954 -- # '[' -z 59354 ']' 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@958 -- # kill -0 59354 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@959 -- # uname 00:08:37.727 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:37.728 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 59354 00:08:37.728 killing process with pid 59354 00:08:37.728 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:37.728 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:37.728 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 59354' 00:08:37.728 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@973 -- # kill 59354 00:08:37.728 [2024-11-20 15:56:35.913627] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:37.728 [2024-11-20 15:56:35.913708] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:37.728 15:56:35 bdev_raid.raid1_resize_test -- common/autotest_common.sh@978 -- # wait 59354 00:08:37.728 [2024-11-20 15:56:35.914133] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:37.728 [2024-11-20 15:56:35.914150] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:08:37.728 [2024-11-20 15:56:35.924598] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:38.661 15:56:36 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:08:38.661 00:08:38.661 real 0m1.782s 00:08:38.661 user 0m1.958s 00:08:38.661 sys 0m0.238s 00:08:38.661 15:56:36 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:38.661 ************************************ 00:08:38.661 END TEST raid1_resize_test 00:08:38.661 ************************************ 00:08:38.661 15:56:36 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.661 15:56:36 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:08:38.661 15:56:36 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:38.661 15:56:36 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 2 false 00:08:38.661 15:56:36 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:38.661 15:56:36 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:38.661 15:56:36 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:38.661 ************************************ 00:08:38.661 START TEST raid_state_function_test 00:08:38.661 ************************************ 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 2 false 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:38.661 Process raid pid: 59411 00:08:38.661 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=59411 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 59411' 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 59411 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 59411 ']' 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.661 15:56:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:38.661 [2024-11-20 15:56:36.744930] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:38.661 [2024-11-20 15:56:36.745047] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:38.661 [2024-11-20 15:56:36.900202] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:38.918 [2024-11-20 15:56:36.997642] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:38.918 [2024-11-20 15:56:37.133478] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:38.918 [2024-11-20 15:56:37.133509] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.481 [2024-11-20 15:56:37.591763] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:39.481 [2024-11-20 15:56:37.591911] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:39.481 [2024-11-20 15:56:37.591977] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:39.481 [2024-11-20 15:56:37.592006] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.481 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.482 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.482 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:39.482 "name": "Existed_Raid", 00:08:39.482 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.482 "strip_size_kb": 64, 00:08:39.482 "state": "configuring", 00:08:39.482 "raid_level": "raid0", 00:08:39.482 "superblock": false, 00:08:39.482 "num_base_bdevs": 2, 00:08:39.482 "num_base_bdevs_discovered": 0, 00:08:39.482 "num_base_bdevs_operational": 2, 00:08:39.482 "base_bdevs_list": [ 00:08:39.482 { 00:08:39.482 "name": "BaseBdev1", 00:08:39.482 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.482 "is_configured": false, 00:08:39.482 "data_offset": 0, 00:08:39.482 "data_size": 0 00:08:39.482 }, 00:08:39.482 { 00:08:39.482 "name": "BaseBdev2", 00:08:39.482 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.482 "is_configured": false, 00:08:39.482 "data_offset": 0, 00:08:39.482 "data_size": 0 00:08:39.482 } 00:08:39.482 ] 00:08:39.482 }' 00:08:39.482 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:39.482 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.739 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:39.739 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.739 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.739 [2024-11-20 15:56:37.903784] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:39.739 [2024-11-20 15:56:37.903910] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:08:39.739 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.739 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:39.739 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.739 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.740 [2024-11-20 15:56:37.911785] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:39.740 [2024-11-20 15:56:37.911901] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:39.740 [2024-11-20 15:56:37.911957] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:39.740 [2024-11-20 15:56:37.911985] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.740 BaseBdev1 00:08:39.740 [2024-11-20 15:56:37.943982] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.740 [ 00:08:39.740 { 00:08:39.740 "name": "BaseBdev1", 00:08:39.740 "aliases": [ 00:08:39.740 "fca65b04-c286-47aa-8903-0c018cc74ede" 00:08:39.740 ], 00:08:39.740 "product_name": "Malloc disk", 00:08:39.740 "block_size": 512, 00:08:39.740 "num_blocks": 65536, 00:08:39.740 "uuid": "fca65b04-c286-47aa-8903-0c018cc74ede", 00:08:39.740 "assigned_rate_limits": { 00:08:39.740 "rw_ios_per_sec": 0, 00:08:39.740 "rw_mbytes_per_sec": 0, 00:08:39.740 "r_mbytes_per_sec": 0, 00:08:39.740 "w_mbytes_per_sec": 0 00:08:39.740 }, 00:08:39.740 "claimed": true, 00:08:39.740 "claim_type": "exclusive_write", 00:08:39.740 "zoned": false, 00:08:39.740 "supported_io_types": { 00:08:39.740 "read": true, 00:08:39.740 "write": true, 00:08:39.740 "unmap": true, 00:08:39.740 "flush": true, 00:08:39.740 "reset": true, 00:08:39.740 "nvme_admin": false, 00:08:39.740 "nvme_io": false, 00:08:39.740 "nvme_io_md": false, 00:08:39.740 "write_zeroes": true, 00:08:39.740 "zcopy": true, 00:08:39.740 "get_zone_info": false, 00:08:39.740 "zone_management": false, 00:08:39.740 "zone_append": false, 00:08:39.740 "compare": false, 00:08:39.740 "compare_and_write": false, 00:08:39.740 "abort": true, 00:08:39.740 "seek_hole": false, 00:08:39.740 "seek_data": false, 00:08:39.740 "copy": true, 00:08:39.740 "nvme_iov_md": false 00:08:39.740 }, 00:08:39.740 "memory_domains": [ 00:08:39.740 { 00:08:39.740 "dma_device_id": "system", 00:08:39.740 "dma_device_type": 1 00:08:39.740 }, 00:08:39.740 { 00:08:39.740 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:39.740 "dma_device_type": 2 00:08:39.740 } 00:08:39.740 ], 00:08:39.740 "driver_specific": {} 00:08:39.740 } 00:08:39.740 ] 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.740 15:56:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.998 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:39.998 "name": "Existed_Raid", 00:08:39.998 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.998 "strip_size_kb": 64, 00:08:39.998 "state": "configuring", 00:08:39.998 "raid_level": "raid0", 00:08:39.998 "superblock": false, 00:08:39.998 "num_base_bdevs": 2, 00:08:39.998 "num_base_bdevs_discovered": 1, 00:08:39.998 "num_base_bdevs_operational": 2, 00:08:39.998 "base_bdevs_list": [ 00:08:39.998 { 00:08:39.998 "name": "BaseBdev1", 00:08:39.998 "uuid": "fca65b04-c286-47aa-8903-0c018cc74ede", 00:08:39.998 "is_configured": true, 00:08:39.998 "data_offset": 0, 00:08:39.998 "data_size": 65536 00:08:39.998 }, 00:08:39.998 { 00:08:39.998 "name": "BaseBdev2", 00:08:39.998 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.998 "is_configured": false, 00:08:39.998 "data_offset": 0, 00:08:39.998 "data_size": 0 00:08:39.998 } 00:08:39.998 ] 00:08:39.998 }' 00:08:39.998 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:39.998 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.255 [2024-11-20 15:56:38.292095] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:40.255 [2024-11-20 15:56:38.292274] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.255 [2024-11-20 15:56:38.300144] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:40.255 [2024-11-20 15:56:38.302047] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:40.255 [2024-11-20 15:56:38.302165] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:40.255 "name": "Existed_Raid", 00:08:40.255 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.255 "strip_size_kb": 64, 00:08:40.255 "state": "configuring", 00:08:40.255 "raid_level": "raid0", 00:08:40.255 "superblock": false, 00:08:40.255 "num_base_bdevs": 2, 00:08:40.255 "num_base_bdevs_discovered": 1, 00:08:40.255 "num_base_bdevs_operational": 2, 00:08:40.255 "base_bdevs_list": [ 00:08:40.255 { 00:08:40.255 "name": "BaseBdev1", 00:08:40.255 "uuid": "fca65b04-c286-47aa-8903-0c018cc74ede", 00:08:40.255 "is_configured": true, 00:08:40.255 "data_offset": 0, 00:08:40.255 "data_size": 65536 00:08:40.255 }, 00:08:40.255 { 00:08:40.255 "name": "BaseBdev2", 00:08:40.255 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.255 "is_configured": false, 00:08:40.255 "data_offset": 0, 00:08:40.255 "data_size": 0 00:08:40.255 } 00:08:40.255 ] 00:08:40.255 }' 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:40.255 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.512 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:40.512 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.512 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.512 [2024-11-20 15:56:38.638447] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:40.512 BaseBdev2 00:08:40.512 [2024-11-20 15:56:38.638647] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:08:40.512 [2024-11-20 15:56:38.638660] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:08:40.513 [2024-11-20 15:56:38.638944] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:40.513 [2024-11-20 15:56:38.639068] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:08:40.513 [2024-11-20 15:56:38.639076] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:08:40.513 [2024-11-20 15:56:38.639263] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.513 [ 00:08:40.513 { 00:08:40.513 "name": "BaseBdev2", 00:08:40.513 "aliases": [ 00:08:40.513 "d4c24e8d-270e-468b-82aa-243fd74c8a76" 00:08:40.513 ], 00:08:40.513 "product_name": "Malloc disk", 00:08:40.513 "block_size": 512, 00:08:40.513 "num_blocks": 65536, 00:08:40.513 "uuid": "d4c24e8d-270e-468b-82aa-243fd74c8a76", 00:08:40.513 "assigned_rate_limits": { 00:08:40.513 "rw_ios_per_sec": 0, 00:08:40.513 "rw_mbytes_per_sec": 0, 00:08:40.513 "r_mbytes_per_sec": 0, 00:08:40.513 "w_mbytes_per_sec": 0 00:08:40.513 }, 00:08:40.513 "claimed": true, 00:08:40.513 "claim_type": "exclusive_write", 00:08:40.513 "zoned": false, 00:08:40.513 "supported_io_types": { 00:08:40.513 "read": true, 00:08:40.513 "write": true, 00:08:40.513 "unmap": true, 00:08:40.513 "flush": true, 00:08:40.513 "reset": true, 00:08:40.513 "nvme_admin": false, 00:08:40.513 "nvme_io": false, 00:08:40.513 "nvme_io_md": false, 00:08:40.513 "write_zeroes": true, 00:08:40.513 "zcopy": true, 00:08:40.513 "get_zone_info": false, 00:08:40.513 "zone_management": false, 00:08:40.513 "zone_append": false, 00:08:40.513 "compare": false, 00:08:40.513 "compare_and_write": false, 00:08:40.513 "abort": true, 00:08:40.513 "seek_hole": false, 00:08:40.513 "seek_data": false, 00:08:40.513 "copy": true, 00:08:40.513 "nvme_iov_md": false 00:08:40.513 }, 00:08:40.513 "memory_domains": [ 00:08:40.513 { 00:08:40.513 "dma_device_id": "system", 00:08:40.513 "dma_device_type": 1 00:08:40.513 }, 00:08:40.513 { 00:08:40.513 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:40.513 "dma_device_type": 2 00:08:40.513 } 00:08:40.513 ], 00:08:40.513 "driver_specific": {} 00:08:40.513 } 00:08:40.513 ] 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:40.513 "name": "Existed_Raid", 00:08:40.513 "uuid": "e732b163-e816-43c3-9d49-a128556d4746", 00:08:40.513 "strip_size_kb": 64, 00:08:40.513 "state": "online", 00:08:40.513 "raid_level": "raid0", 00:08:40.513 "superblock": false, 00:08:40.513 "num_base_bdevs": 2, 00:08:40.513 "num_base_bdevs_discovered": 2, 00:08:40.513 "num_base_bdevs_operational": 2, 00:08:40.513 "base_bdevs_list": [ 00:08:40.513 { 00:08:40.513 "name": "BaseBdev1", 00:08:40.513 "uuid": "fca65b04-c286-47aa-8903-0c018cc74ede", 00:08:40.513 "is_configured": true, 00:08:40.513 "data_offset": 0, 00:08:40.513 "data_size": 65536 00:08:40.513 }, 00:08:40.513 { 00:08:40.513 "name": "BaseBdev2", 00:08:40.513 "uuid": "d4c24e8d-270e-468b-82aa-243fd74c8a76", 00:08:40.513 "is_configured": true, 00:08:40.513 "data_offset": 0, 00:08:40.513 "data_size": 65536 00:08:40.513 } 00:08:40.513 ] 00:08:40.513 }' 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:40.513 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.770 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:40.770 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:40.770 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:40.770 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:40.770 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:40.770 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:40.770 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:40.770 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.770 15:56:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:40.770 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.770 [2024-11-20 15:56:38.974782] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:40.770 15:56:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.770 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:40.770 "name": "Existed_Raid", 00:08:40.770 "aliases": [ 00:08:40.770 "e732b163-e816-43c3-9d49-a128556d4746" 00:08:40.770 ], 00:08:40.770 "product_name": "Raid Volume", 00:08:40.770 "block_size": 512, 00:08:40.770 "num_blocks": 131072, 00:08:40.770 "uuid": "e732b163-e816-43c3-9d49-a128556d4746", 00:08:40.770 "assigned_rate_limits": { 00:08:40.770 "rw_ios_per_sec": 0, 00:08:40.770 "rw_mbytes_per_sec": 0, 00:08:40.770 "r_mbytes_per_sec": 0, 00:08:40.770 "w_mbytes_per_sec": 0 00:08:40.770 }, 00:08:40.770 "claimed": false, 00:08:40.770 "zoned": false, 00:08:40.770 "supported_io_types": { 00:08:40.770 "read": true, 00:08:40.770 "write": true, 00:08:40.770 "unmap": true, 00:08:40.770 "flush": true, 00:08:40.770 "reset": true, 00:08:40.770 "nvme_admin": false, 00:08:40.770 "nvme_io": false, 00:08:40.770 "nvme_io_md": false, 00:08:40.770 "write_zeroes": true, 00:08:40.770 "zcopy": false, 00:08:40.770 "get_zone_info": false, 00:08:40.770 "zone_management": false, 00:08:40.770 "zone_append": false, 00:08:40.770 "compare": false, 00:08:40.770 "compare_and_write": false, 00:08:40.770 "abort": false, 00:08:40.770 "seek_hole": false, 00:08:40.770 "seek_data": false, 00:08:40.770 "copy": false, 00:08:40.770 "nvme_iov_md": false 00:08:40.770 }, 00:08:40.770 "memory_domains": [ 00:08:40.770 { 00:08:40.770 "dma_device_id": "system", 00:08:40.770 "dma_device_type": 1 00:08:40.770 }, 00:08:40.770 { 00:08:40.770 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:40.770 "dma_device_type": 2 00:08:40.770 }, 00:08:40.770 { 00:08:40.770 "dma_device_id": "system", 00:08:40.770 "dma_device_type": 1 00:08:40.770 }, 00:08:40.770 { 00:08:40.770 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:40.770 "dma_device_type": 2 00:08:40.770 } 00:08:40.770 ], 00:08:40.770 "driver_specific": { 00:08:40.770 "raid": { 00:08:40.770 "uuid": "e732b163-e816-43c3-9d49-a128556d4746", 00:08:40.770 "strip_size_kb": 64, 00:08:40.770 "state": "online", 00:08:40.770 "raid_level": "raid0", 00:08:40.770 "superblock": false, 00:08:40.770 "num_base_bdevs": 2, 00:08:40.770 "num_base_bdevs_discovered": 2, 00:08:40.770 "num_base_bdevs_operational": 2, 00:08:40.770 "base_bdevs_list": [ 00:08:40.770 { 00:08:40.770 "name": "BaseBdev1", 00:08:40.770 "uuid": "fca65b04-c286-47aa-8903-0c018cc74ede", 00:08:40.770 "is_configured": true, 00:08:40.770 "data_offset": 0, 00:08:40.770 "data_size": 65536 00:08:40.770 }, 00:08:40.770 { 00:08:40.770 "name": "BaseBdev2", 00:08:40.770 "uuid": "d4c24e8d-270e-468b-82aa-243fd74c8a76", 00:08:40.770 "is_configured": true, 00:08:40.770 "data_offset": 0, 00:08:40.770 "data_size": 65536 00:08:40.770 } 00:08:40.770 ] 00:08:40.770 } 00:08:40.770 } 00:08:40.770 }' 00:08:40.770 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:41.027 BaseBdev2' 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.027 [2024-11-20 15:56:39.154623] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:41.027 [2024-11-20 15:56:39.154743] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:41.027 [2024-11-20 15:56:39.154826] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.027 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.028 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:41.028 "name": "Existed_Raid", 00:08:41.028 "uuid": "e732b163-e816-43c3-9d49-a128556d4746", 00:08:41.028 "strip_size_kb": 64, 00:08:41.028 "state": "offline", 00:08:41.028 "raid_level": "raid0", 00:08:41.028 "superblock": false, 00:08:41.028 "num_base_bdevs": 2, 00:08:41.028 "num_base_bdevs_discovered": 1, 00:08:41.028 "num_base_bdevs_operational": 1, 00:08:41.028 "base_bdevs_list": [ 00:08:41.028 { 00:08:41.028 "name": null, 00:08:41.028 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:41.028 "is_configured": false, 00:08:41.028 "data_offset": 0, 00:08:41.028 "data_size": 65536 00:08:41.028 }, 00:08:41.028 { 00:08:41.028 "name": "BaseBdev2", 00:08:41.028 "uuid": "d4c24e8d-270e-468b-82aa-243fd74c8a76", 00:08:41.028 "is_configured": true, 00:08:41.028 "data_offset": 0, 00:08:41.028 "data_size": 65536 00:08:41.028 } 00:08:41.028 ] 00:08:41.028 }' 00:08:41.028 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:41.028 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.285 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:41.285 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:41.285 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:41.285 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.285 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.285 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:41.285 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.542 [2024-11-20 15:56:39.549388] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:41.542 [2024-11-20 15:56:39.549520] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 59411 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 59411 ']' 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 59411 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 59411 00:08:41.542 killing process with pid 59411 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 59411' 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 59411 00:08:41.542 [2024-11-20 15:56:39.660913] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:41.542 15:56:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 59411 00:08:41.542 [2024-11-20 15:56:39.669259] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:42.105 ************************************ 00:08:42.105 END TEST raid_state_function_test 00:08:42.105 ************************************ 00:08:42.105 15:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:42.105 00:08:42.105 real 0m3.561s 00:08:42.105 user 0m5.253s 00:08:42.105 sys 0m0.528s 00:08:42.105 15:56:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:42.105 15:56:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.105 15:56:40 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 2 true 00:08:42.105 15:56:40 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:42.105 15:56:40 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:42.105 15:56:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:42.105 ************************************ 00:08:42.105 START TEST raid_state_function_test_sb 00:08:42.106 ************************************ 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 2 true 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:42.106 Process raid pid: 59642 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=59642 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 59642' 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 59642 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 59642 ']' 00:08:42.106 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.106 15:56:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:42.106 [2024-11-20 15:56:40.345834] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:42.106 [2024-11-20 15:56:40.345968] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:42.362 [2024-11-20 15:56:40.504895] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:42.362 [2024-11-20 15:56:40.605722] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:42.620 [2024-11-20 15:56:40.742739] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:42.620 [2024-11-20 15:56:40.742901] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:43.185 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:43.185 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:08:43.185 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:43.185 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.185 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.185 [2024-11-20 15:56:41.236493] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:43.185 [2024-11-20 15:56:41.236648] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:43.185 [2024-11-20 15:56:41.236738] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:43.185 [2024-11-20 15:56:41.236768] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:43.185 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.185 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:08:43.185 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:43.185 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:43.185 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:43.185 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:43.185 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:43.186 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:43.186 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:43.186 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:43.186 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:43.186 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:43.186 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.186 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.186 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:43.186 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.186 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:43.186 "name": "Existed_Raid", 00:08:43.186 "uuid": "7696fef3-9b28-4a17-be30-d4312c0eabd0", 00:08:43.186 "strip_size_kb": 64, 00:08:43.186 "state": "configuring", 00:08:43.186 "raid_level": "raid0", 00:08:43.186 "superblock": true, 00:08:43.186 "num_base_bdevs": 2, 00:08:43.186 "num_base_bdevs_discovered": 0, 00:08:43.186 "num_base_bdevs_operational": 2, 00:08:43.186 "base_bdevs_list": [ 00:08:43.186 { 00:08:43.186 "name": "BaseBdev1", 00:08:43.186 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:43.186 "is_configured": false, 00:08:43.186 "data_offset": 0, 00:08:43.186 "data_size": 0 00:08:43.186 }, 00:08:43.186 { 00:08:43.186 "name": "BaseBdev2", 00:08:43.186 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:43.186 "is_configured": false, 00:08:43.186 "data_offset": 0, 00:08:43.186 "data_size": 0 00:08:43.186 } 00:08:43.186 ] 00:08:43.186 }' 00:08:43.186 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:43.186 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.444 [2024-11-20 15:56:41.552517] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:43.444 [2024-11-20 15:56:41.552648] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.444 [2024-11-20 15:56:41.560518] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:43.444 [2024-11-20 15:56:41.560637] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:43.444 [2024-11-20 15:56:41.560706] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:43.444 [2024-11-20 15:56:41.560737] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.444 [2024-11-20 15:56:41.592775] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:43.444 BaseBdev1 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.444 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.444 [ 00:08:43.444 { 00:08:43.444 "name": "BaseBdev1", 00:08:43.444 "aliases": [ 00:08:43.444 "6cb66b5c-659c-48ff-af59-ba2863128bcc" 00:08:43.444 ], 00:08:43.444 "product_name": "Malloc disk", 00:08:43.444 "block_size": 512, 00:08:43.444 "num_blocks": 65536, 00:08:43.444 "uuid": "6cb66b5c-659c-48ff-af59-ba2863128bcc", 00:08:43.444 "assigned_rate_limits": { 00:08:43.444 "rw_ios_per_sec": 0, 00:08:43.444 "rw_mbytes_per_sec": 0, 00:08:43.444 "r_mbytes_per_sec": 0, 00:08:43.444 "w_mbytes_per_sec": 0 00:08:43.444 }, 00:08:43.444 "claimed": true, 00:08:43.444 "claim_type": "exclusive_write", 00:08:43.444 "zoned": false, 00:08:43.444 "supported_io_types": { 00:08:43.444 "read": true, 00:08:43.444 "write": true, 00:08:43.444 "unmap": true, 00:08:43.444 "flush": true, 00:08:43.444 "reset": true, 00:08:43.444 "nvme_admin": false, 00:08:43.444 "nvme_io": false, 00:08:43.444 "nvme_io_md": false, 00:08:43.444 "write_zeroes": true, 00:08:43.444 "zcopy": true, 00:08:43.444 "get_zone_info": false, 00:08:43.444 "zone_management": false, 00:08:43.444 "zone_append": false, 00:08:43.444 "compare": false, 00:08:43.444 "compare_and_write": false, 00:08:43.444 "abort": true, 00:08:43.444 "seek_hole": false, 00:08:43.444 "seek_data": false, 00:08:43.444 "copy": true, 00:08:43.444 "nvme_iov_md": false 00:08:43.444 }, 00:08:43.444 "memory_domains": [ 00:08:43.444 { 00:08:43.444 "dma_device_id": "system", 00:08:43.444 "dma_device_type": 1 00:08:43.444 }, 00:08:43.444 { 00:08:43.444 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:43.444 "dma_device_type": 2 00:08:43.444 } 00:08:43.444 ], 00:08:43.444 "driver_specific": {} 00:08:43.444 } 00:08:43.444 ] 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:43.445 "name": "Existed_Raid", 00:08:43.445 "uuid": "dfb18b08-fca9-4c99-83fc-23e6da3300e1", 00:08:43.445 "strip_size_kb": 64, 00:08:43.445 "state": "configuring", 00:08:43.445 "raid_level": "raid0", 00:08:43.445 "superblock": true, 00:08:43.445 "num_base_bdevs": 2, 00:08:43.445 "num_base_bdevs_discovered": 1, 00:08:43.445 "num_base_bdevs_operational": 2, 00:08:43.445 "base_bdevs_list": [ 00:08:43.445 { 00:08:43.445 "name": "BaseBdev1", 00:08:43.445 "uuid": "6cb66b5c-659c-48ff-af59-ba2863128bcc", 00:08:43.445 "is_configured": true, 00:08:43.445 "data_offset": 2048, 00:08:43.445 "data_size": 63488 00:08:43.445 }, 00:08:43.445 { 00:08:43.445 "name": "BaseBdev2", 00:08:43.445 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:43.445 "is_configured": false, 00:08:43.445 "data_offset": 0, 00:08:43.445 "data_size": 0 00:08:43.445 } 00:08:43.445 ] 00:08:43.445 }' 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:43.445 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.789 [2024-11-20 15:56:41.908918] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:43.789 [2024-11-20 15:56:41.909065] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.789 [2024-11-20 15:56:41.916966] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:43.789 [2024-11-20 15:56:41.918816] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:43.789 [2024-11-20 15:56:41.918946] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:43.789 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:43.790 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.790 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.790 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:43.790 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.790 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:43.790 "name": "Existed_Raid", 00:08:43.790 "uuid": "0fb62c9f-cc4b-4805-a91c-2dba12bfd527", 00:08:43.790 "strip_size_kb": 64, 00:08:43.790 "state": "configuring", 00:08:43.790 "raid_level": "raid0", 00:08:43.790 "superblock": true, 00:08:43.790 "num_base_bdevs": 2, 00:08:43.790 "num_base_bdevs_discovered": 1, 00:08:43.790 "num_base_bdevs_operational": 2, 00:08:43.790 "base_bdevs_list": [ 00:08:43.790 { 00:08:43.790 "name": "BaseBdev1", 00:08:43.790 "uuid": "6cb66b5c-659c-48ff-af59-ba2863128bcc", 00:08:43.790 "is_configured": true, 00:08:43.790 "data_offset": 2048, 00:08:43.790 "data_size": 63488 00:08:43.790 }, 00:08:43.790 { 00:08:43.790 "name": "BaseBdev2", 00:08:43.790 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:43.790 "is_configured": false, 00:08:43.790 "data_offset": 0, 00:08:43.790 "data_size": 0 00:08:43.790 } 00:08:43.790 ] 00:08:43.790 }' 00:08:43.790 15:56:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:43.790 15:56:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.050 [2024-11-20 15:56:42.251440] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:44.050 BaseBdev2 00:08:44.050 [2024-11-20 15:56:42.251867] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:08:44.050 [2024-11-20 15:56:42.251886] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:44.050 [2024-11-20 15:56:42.252144] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:44.050 [2024-11-20 15:56:42.252277] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:08:44.050 [2024-11-20 15:56:42.252289] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.050 [2024-11-20 15:56:42.252412] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.050 [ 00:08:44.050 { 00:08:44.050 "name": "BaseBdev2", 00:08:44.050 "aliases": [ 00:08:44.050 "8a966b29-37a3-4a84-9ac0-437235bcaaf8" 00:08:44.050 ], 00:08:44.050 "product_name": "Malloc disk", 00:08:44.050 "block_size": 512, 00:08:44.050 "num_blocks": 65536, 00:08:44.050 "uuid": "8a966b29-37a3-4a84-9ac0-437235bcaaf8", 00:08:44.050 "assigned_rate_limits": { 00:08:44.050 "rw_ios_per_sec": 0, 00:08:44.050 "rw_mbytes_per_sec": 0, 00:08:44.050 "r_mbytes_per_sec": 0, 00:08:44.050 "w_mbytes_per_sec": 0 00:08:44.050 }, 00:08:44.050 "claimed": true, 00:08:44.050 "claim_type": "exclusive_write", 00:08:44.050 "zoned": false, 00:08:44.050 "supported_io_types": { 00:08:44.050 "read": true, 00:08:44.050 "write": true, 00:08:44.050 "unmap": true, 00:08:44.050 "flush": true, 00:08:44.050 "reset": true, 00:08:44.050 "nvme_admin": false, 00:08:44.050 "nvme_io": false, 00:08:44.050 "nvme_io_md": false, 00:08:44.050 "write_zeroes": true, 00:08:44.050 "zcopy": true, 00:08:44.050 "get_zone_info": false, 00:08:44.050 "zone_management": false, 00:08:44.050 "zone_append": false, 00:08:44.050 "compare": false, 00:08:44.050 "compare_and_write": false, 00:08:44.050 "abort": true, 00:08:44.050 "seek_hole": false, 00:08:44.050 "seek_data": false, 00:08:44.050 "copy": true, 00:08:44.050 "nvme_iov_md": false 00:08:44.050 }, 00:08:44.050 "memory_domains": [ 00:08:44.050 { 00:08:44.050 "dma_device_id": "system", 00:08:44.050 "dma_device_type": 1 00:08:44.050 }, 00:08:44.050 { 00:08:44.050 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:44.050 "dma_device_type": 2 00:08:44.050 } 00:08:44.050 ], 00:08:44.050 "driver_specific": {} 00:08:44.050 } 00:08:44.050 ] 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.050 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.309 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:44.309 "name": "Existed_Raid", 00:08:44.309 "uuid": "0fb62c9f-cc4b-4805-a91c-2dba12bfd527", 00:08:44.309 "strip_size_kb": 64, 00:08:44.309 "state": "online", 00:08:44.309 "raid_level": "raid0", 00:08:44.309 "superblock": true, 00:08:44.309 "num_base_bdevs": 2, 00:08:44.309 "num_base_bdevs_discovered": 2, 00:08:44.309 "num_base_bdevs_operational": 2, 00:08:44.309 "base_bdevs_list": [ 00:08:44.309 { 00:08:44.309 "name": "BaseBdev1", 00:08:44.309 "uuid": "6cb66b5c-659c-48ff-af59-ba2863128bcc", 00:08:44.309 "is_configured": true, 00:08:44.309 "data_offset": 2048, 00:08:44.309 "data_size": 63488 00:08:44.309 }, 00:08:44.309 { 00:08:44.309 "name": "BaseBdev2", 00:08:44.309 "uuid": "8a966b29-37a3-4a84-9ac0-437235bcaaf8", 00:08:44.309 "is_configured": true, 00:08:44.309 "data_offset": 2048, 00:08:44.309 "data_size": 63488 00:08:44.309 } 00:08:44.309 ] 00:08:44.309 }' 00:08:44.309 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:44.309 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.569 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:44.569 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:44.569 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:44.569 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:44.569 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:44.569 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:44.569 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:44.569 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:44.569 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.569 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.569 [2024-11-20 15:56:42.579870] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:44.569 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.569 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:44.569 "name": "Existed_Raid", 00:08:44.569 "aliases": [ 00:08:44.569 "0fb62c9f-cc4b-4805-a91c-2dba12bfd527" 00:08:44.569 ], 00:08:44.569 "product_name": "Raid Volume", 00:08:44.569 "block_size": 512, 00:08:44.569 "num_blocks": 126976, 00:08:44.569 "uuid": "0fb62c9f-cc4b-4805-a91c-2dba12bfd527", 00:08:44.569 "assigned_rate_limits": { 00:08:44.569 "rw_ios_per_sec": 0, 00:08:44.569 "rw_mbytes_per_sec": 0, 00:08:44.569 "r_mbytes_per_sec": 0, 00:08:44.569 "w_mbytes_per_sec": 0 00:08:44.569 }, 00:08:44.569 "claimed": false, 00:08:44.569 "zoned": false, 00:08:44.569 "supported_io_types": { 00:08:44.569 "read": true, 00:08:44.569 "write": true, 00:08:44.569 "unmap": true, 00:08:44.569 "flush": true, 00:08:44.569 "reset": true, 00:08:44.569 "nvme_admin": false, 00:08:44.569 "nvme_io": false, 00:08:44.569 "nvme_io_md": false, 00:08:44.569 "write_zeroes": true, 00:08:44.569 "zcopy": false, 00:08:44.569 "get_zone_info": false, 00:08:44.569 "zone_management": false, 00:08:44.569 "zone_append": false, 00:08:44.569 "compare": false, 00:08:44.569 "compare_and_write": false, 00:08:44.569 "abort": false, 00:08:44.569 "seek_hole": false, 00:08:44.569 "seek_data": false, 00:08:44.569 "copy": false, 00:08:44.569 "nvme_iov_md": false 00:08:44.569 }, 00:08:44.569 "memory_domains": [ 00:08:44.569 { 00:08:44.569 "dma_device_id": "system", 00:08:44.569 "dma_device_type": 1 00:08:44.569 }, 00:08:44.569 { 00:08:44.569 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:44.569 "dma_device_type": 2 00:08:44.569 }, 00:08:44.569 { 00:08:44.569 "dma_device_id": "system", 00:08:44.569 "dma_device_type": 1 00:08:44.569 }, 00:08:44.569 { 00:08:44.569 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:44.569 "dma_device_type": 2 00:08:44.569 } 00:08:44.569 ], 00:08:44.569 "driver_specific": { 00:08:44.569 "raid": { 00:08:44.569 "uuid": "0fb62c9f-cc4b-4805-a91c-2dba12bfd527", 00:08:44.569 "strip_size_kb": 64, 00:08:44.569 "state": "online", 00:08:44.569 "raid_level": "raid0", 00:08:44.569 "superblock": true, 00:08:44.569 "num_base_bdevs": 2, 00:08:44.569 "num_base_bdevs_discovered": 2, 00:08:44.569 "num_base_bdevs_operational": 2, 00:08:44.569 "base_bdevs_list": [ 00:08:44.569 { 00:08:44.569 "name": "BaseBdev1", 00:08:44.569 "uuid": "6cb66b5c-659c-48ff-af59-ba2863128bcc", 00:08:44.569 "is_configured": true, 00:08:44.569 "data_offset": 2048, 00:08:44.569 "data_size": 63488 00:08:44.569 }, 00:08:44.569 { 00:08:44.569 "name": "BaseBdev2", 00:08:44.569 "uuid": "8a966b29-37a3-4a84-9ac0-437235bcaaf8", 00:08:44.569 "is_configured": true, 00:08:44.569 "data_offset": 2048, 00:08:44.569 "data_size": 63488 00:08:44.569 } 00:08:44.569 ] 00:08:44.569 } 00:08:44.569 } 00:08:44.569 }' 00:08:44.569 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:44.569 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:44.569 BaseBdev2' 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.570 [2024-11-20 15:56:42.743637] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:44.570 [2024-11-20 15:56:42.743790] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:44.570 [2024-11-20 15:56:42.743901] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.570 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.829 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.829 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:44.829 "name": "Existed_Raid", 00:08:44.829 "uuid": "0fb62c9f-cc4b-4805-a91c-2dba12bfd527", 00:08:44.829 "strip_size_kb": 64, 00:08:44.829 "state": "offline", 00:08:44.829 "raid_level": "raid0", 00:08:44.829 "superblock": true, 00:08:44.829 "num_base_bdevs": 2, 00:08:44.829 "num_base_bdevs_discovered": 1, 00:08:44.829 "num_base_bdevs_operational": 1, 00:08:44.829 "base_bdevs_list": [ 00:08:44.829 { 00:08:44.829 "name": null, 00:08:44.829 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:44.829 "is_configured": false, 00:08:44.829 "data_offset": 0, 00:08:44.829 "data_size": 63488 00:08:44.829 }, 00:08:44.829 { 00:08:44.829 "name": "BaseBdev2", 00:08:44.829 "uuid": "8a966b29-37a3-4a84-9ac0-437235bcaaf8", 00:08:44.829 "is_configured": true, 00:08:44.829 "data_offset": 2048, 00:08:44.829 "data_size": 63488 00:08:44.829 } 00:08:44.829 ] 00:08:44.829 }' 00:08:44.829 15:56:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:44.829 15:56:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:45.086 15:56:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:45.086 15:56:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:45.086 15:56:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.086 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.086 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:45.086 15:56:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:45.086 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.086 15:56:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:45.086 15:56:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:45.086 15:56:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:45.087 [2024-11-20 15:56:43.137981] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:45.087 [2024-11-20 15:56:43.138114] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 59642 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 59642 ']' 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 59642 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 59642 00:08:45.087 killing process with pid 59642 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 59642' 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 59642 00:08:45.087 [2024-11-20 15:56:43.241401] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:45.087 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 59642 00:08:45.087 [2024-11-20 15:56:43.249720] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:45.653 ************************************ 00:08:45.653 END TEST raid_state_function_test_sb 00:08:45.653 ************************************ 00:08:45.653 15:56:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:45.653 00:08:45.653 real 0m3.546s 00:08:45.653 user 0m5.199s 00:08:45.653 sys 0m0.536s 00:08:45.653 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:45.653 15:56:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:45.653 15:56:43 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 2 00:08:45.653 15:56:43 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:08:45.653 15:56:43 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:45.653 15:56:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:45.653 ************************************ 00:08:45.653 START TEST raid_superblock_test 00:08:45.653 ************************************ 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 2 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=59883 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 59883 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 59883 ']' 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:45.653 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:45.653 15:56:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.911 [2024-11-20 15:56:43.928617] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:45.911 [2024-11-20 15:56:43.928948] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid59883 ] 00:08:45.911 [2024-11-20 15:56:44.084273] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:46.169 [2024-11-20 15:56:44.183451] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:46.169 [2024-11-20 15:56:44.319262] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:46.169 [2024-11-20 15:56:44.319303] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.733 malloc1 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.733 [2024-11-20 15:56:44.834381] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:46.733 [2024-11-20 15:56:44.834568] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:46.733 [2024-11-20 15:56:44.834610] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:46.733 [2024-11-20 15:56:44.834989] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:46.733 [2024-11-20 15:56:44.837189] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:46.733 [2024-11-20 15:56:44.837314] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:46.733 pt1 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:46.733 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.734 malloc2 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.734 [2024-11-20 15:56:44.878328] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:46.734 [2024-11-20 15:56:44.878491] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:46.734 [2024-11-20 15:56:44.878521] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:46.734 [2024-11-20 15:56:44.878531] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:46.734 [2024-11-20 15:56:44.880759] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:46.734 [2024-11-20 15:56:44.880860] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:46.734 pt2 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.734 [2024-11-20 15:56:44.890399] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:46.734 [2024-11-20 15:56:44.892337] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:46.734 [2024-11-20 15:56:44.892574] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:08:46.734 [2024-11-20 15:56:44.892645] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:46.734 [2024-11-20 15:56:44.892952] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:46.734 [2024-11-20 15:56:44.893157] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:08:46.734 [2024-11-20 15:56:44.893172] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:08:46.734 [2024-11-20 15:56:44.893324] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:46.734 "name": "raid_bdev1", 00:08:46.734 "uuid": "8afdb292-a0db-4a39-8de0-cdd529e32bf9", 00:08:46.734 "strip_size_kb": 64, 00:08:46.734 "state": "online", 00:08:46.734 "raid_level": "raid0", 00:08:46.734 "superblock": true, 00:08:46.734 "num_base_bdevs": 2, 00:08:46.734 "num_base_bdevs_discovered": 2, 00:08:46.734 "num_base_bdevs_operational": 2, 00:08:46.734 "base_bdevs_list": [ 00:08:46.734 { 00:08:46.734 "name": "pt1", 00:08:46.734 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:46.734 "is_configured": true, 00:08:46.734 "data_offset": 2048, 00:08:46.734 "data_size": 63488 00:08:46.734 }, 00:08:46.734 { 00:08:46.734 "name": "pt2", 00:08:46.734 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:46.734 "is_configured": true, 00:08:46.734 "data_offset": 2048, 00:08:46.734 "data_size": 63488 00:08:46.734 } 00:08:46.734 ] 00:08:46.734 }' 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:46.734 15:56:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.991 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:46.991 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:46.991 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:46.991 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:46.991 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:46.991 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:46.991 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:46.991 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:46.991 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.991 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.991 [2024-11-20 15:56:45.214901] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:46.991 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.991 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:46.991 "name": "raid_bdev1", 00:08:46.991 "aliases": [ 00:08:46.991 "8afdb292-a0db-4a39-8de0-cdd529e32bf9" 00:08:46.991 ], 00:08:46.991 "product_name": "Raid Volume", 00:08:46.991 "block_size": 512, 00:08:46.991 "num_blocks": 126976, 00:08:46.991 "uuid": "8afdb292-a0db-4a39-8de0-cdd529e32bf9", 00:08:46.991 "assigned_rate_limits": { 00:08:46.991 "rw_ios_per_sec": 0, 00:08:46.991 "rw_mbytes_per_sec": 0, 00:08:46.991 "r_mbytes_per_sec": 0, 00:08:46.991 "w_mbytes_per_sec": 0 00:08:46.991 }, 00:08:46.991 "claimed": false, 00:08:46.991 "zoned": false, 00:08:46.991 "supported_io_types": { 00:08:46.991 "read": true, 00:08:46.991 "write": true, 00:08:46.991 "unmap": true, 00:08:46.991 "flush": true, 00:08:46.991 "reset": true, 00:08:46.991 "nvme_admin": false, 00:08:46.991 "nvme_io": false, 00:08:46.991 "nvme_io_md": false, 00:08:46.991 "write_zeroes": true, 00:08:46.991 "zcopy": false, 00:08:46.991 "get_zone_info": false, 00:08:46.991 "zone_management": false, 00:08:46.992 "zone_append": false, 00:08:46.992 "compare": false, 00:08:46.992 "compare_and_write": false, 00:08:46.992 "abort": false, 00:08:46.992 "seek_hole": false, 00:08:46.992 "seek_data": false, 00:08:46.992 "copy": false, 00:08:46.992 "nvme_iov_md": false 00:08:46.992 }, 00:08:46.992 "memory_domains": [ 00:08:46.992 { 00:08:46.992 "dma_device_id": "system", 00:08:46.992 "dma_device_type": 1 00:08:46.992 }, 00:08:46.992 { 00:08:46.992 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:46.992 "dma_device_type": 2 00:08:46.992 }, 00:08:46.992 { 00:08:46.992 "dma_device_id": "system", 00:08:46.992 "dma_device_type": 1 00:08:46.992 }, 00:08:46.992 { 00:08:46.992 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:46.992 "dma_device_type": 2 00:08:46.992 } 00:08:46.992 ], 00:08:46.992 "driver_specific": { 00:08:46.992 "raid": { 00:08:46.992 "uuid": "8afdb292-a0db-4a39-8de0-cdd529e32bf9", 00:08:46.992 "strip_size_kb": 64, 00:08:46.992 "state": "online", 00:08:46.992 "raid_level": "raid0", 00:08:46.992 "superblock": true, 00:08:46.992 "num_base_bdevs": 2, 00:08:46.992 "num_base_bdevs_discovered": 2, 00:08:46.992 "num_base_bdevs_operational": 2, 00:08:46.992 "base_bdevs_list": [ 00:08:46.992 { 00:08:46.992 "name": "pt1", 00:08:46.992 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:46.992 "is_configured": true, 00:08:46.992 "data_offset": 2048, 00:08:46.992 "data_size": 63488 00:08:46.992 }, 00:08:46.992 { 00:08:46.992 "name": "pt2", 00:08:46.992 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:46.992 "is_configured": true, 00:08:46.992 "data_offset": 2048, 00:08:46.992 "data_size": 63488 00:08:46.992 } 00:08:46.992 ] 00:08:46.992 } 00:08:46.992 } 00:08:46.992 }' 00:08:46.992 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:47.301 pt2' 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.301 [2024-11-20 15:56:45.391014] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=8afdb292-a0db-4a39-8de0-cdd529e32bf9 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 8afdb292-a0db-4a39-8de0-cdd529e32bf9 ']' 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.301 [2024-11-20 15:56:45.422491] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:47.301 [2024-11-20 15:56:45.422615] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:47.301 [2024-11-20 15:56:45.422755] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:47.301 [2024-11-20 15:56:45.422857] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:47.301 [2024-11-20 15:56:45.422988] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:47.301 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.302 [2024-11-20 15:56:45.514564] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:47.302 [2024-11-20 15:56:45.516621] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:47.302 [2024-11-20 15:56:45.516710] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:47.302 [2024-11-20 15:56:45.516760] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:47.302 [2024-11-20 15:56:45.516775] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:47.302 [2024-11-20 15:56:45.516789] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:08:47.302 request: 00:08:47.302 { 00:08:47.302 "name": "raid_bdev1", 00:08:47.302 "raid_level": "raid0", 00:08:47.302 "base_bdevs": [ 00:08:47.302 "malloc1", 00:08:47.302 "malloc2" 00:08:47.302 ], 00:08:47.302 "strip_size_kb": 64, 00:08:47.302 "superblock": false, 00:08:47.302 "method": "bdev_raid_create", 00:08:47.302 "req_id": 1 00:08:47.302 } 00:08:47.302 Got JSON-RPC error response 00:08:47.302 response: 00:08:47.302 { 00:08:47.302 "code": -17, 00:08:47.302 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:47.302 } 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.302 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.562 [2024-11-20 15:56:45.558544] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:47.562 [2024-11-20 15:56:45.558715] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:47.562 [2024-11-20 15:56:45.558736] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:47.562 [2024-11-20 15:56:45.558747] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:47.562 [2024-11-20 15:56:45.560928] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:47.562 pt1 00:08:47.562 [2024-11-20 15:56:45.561035] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:47.562 [2024-11-20 15:56:45.561117] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:47.562 [2024-11-20 15:56:45.561166] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 2 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:47.562 "name": "raid_bdev1", 00:08:47.562 "uuid": "8afdb292-a0db-4a39-8de0-cdd529e32bf9", 00:08:47.562 "strip_size_kb": 64, 00:08:47.562 "state": "configuring", 00:08:47.562 "raid_level": "raid0", 00:08:47.562 "superblock": true, 00:08:47.562 "num_base_bdevs": 2, 00:08:47.562 "num_base_bdevs_discovered": 1, 00:08:47.562 "num_base_bdevs_operational": 2, 00:08:47.562 "base_bdevs_list": [ 00:08:47.562 { 00:08:47.562 "name": "pt1", 00:08:47.562 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:47.562 "is_configured": true, 00:08:47.562 "data_offset": 2048, 00:08:47.562 "data_size": 63488 00:08:47.562 }, 00:08:47.562 { 00:08:47.562 "name": null, 00:08:47.562 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:47.562 "is_configured": false, 00:08:47.562 "data_offset": 2048, 00:08:47.562 "data_size": 63488 00:08:47.562 } 00:08:47.562 ] 00:08:47.562 }' 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:47.562 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.819 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.820 [2024-11-20 15:56:45.886624] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:47.820 [2024-11-20 15:56:45.886782] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:47.820 [2024-11-20 15:56:45.886819] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:08:47.820 [2024-11-20 15:56:45.886876] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:47.820 [2024-11-20 15:56:45.887329] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:47.820 [2024-11-20 15:56:45.887426] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:47.820 [2024-11-20 15:56:45.887502] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:47.820 [2024-11-20 15:56:45.887528] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:47.820 [2024-11-20 15:56:45.887652] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:08:47.820 [2024-11-20 15:56:45.887663] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:47.820 [2024-11-20 15:56:45.887909] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:08:47.820 [2024-11-20 15:56:45.888029] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:08:47.820 [2024-11-20 15:56:45.888037] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:08:47.820 [2024-11-20 15:56:45.888157] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:47.820 pt2 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:47.820 "name": "raid_bdev1", 00:08:47.820 "uuid": "8afdb292-a0db-4a39-8de0-cdd529e32bf9", 00:08:47.820 "strip_size_kb": 64, 00:08:47.820 "state": "online", 00:08:47.820 "raid_level": "raid0", 00:08:47.820 "superblock": true, 00:08:47.820 "num_base_bdevs": 2, 00:08:47.820 "num_base_bdevs_discovered": 2, 00:08:47.820 "num_base_bdevs_operational": 2, 00:08:47.820 "base_bdevs_list": [ 00:08:47.820 { 00:08:47.820 "name": "pt1", 00:08:47.820 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:47.820 "is_configured": true, 00:08:47.820 "data_offset": 2048, 00:08:47.820 "data_size": 63488 00:08:47.820 }, 00:08:47.820 { 00:08:47.820 "name": "pt2", 00:08:47.820 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:47.820 "is_configured": true, 00:08:47.820 "data_offset": 2048, 00:08:47.820 "data_size": 63488 00:08:47.820 } 00:08:47.820 ] 00:08:47.820 }' 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:47.820 15:56:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.078 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:48.078 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:48.078 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:48.078 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:48.078 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:48.078 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:48.078 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:48.078 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:48.078 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.078 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.078 [2024-11-20 15:56:46.218956] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:48.078 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.078 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:48.078 "name": "raid_bdev1", 00:08:48.078 "aliases": [ 00:08:48.078 "8afdb292-a0db-4a39-8de0-cdd529e32bf9" 00:08:48.078 ], 00:08:48.078 "product_name": "Raid Volume", 00:08:48.078 "block_size": 512, 00:08:48.078 "num_blocks": 126976, 00:08:48.078 "uuid": "8afdb292-a0db-4a39-8de0-cdd529e32bf9", 00:08:48.078 "assigned_rate_limits": { 00:08:48.078 "rw_ios_per_sec": 0, 00:08:48.078 "rw_mbytes_per_sec": 0, 00:08:48.078 "r_mbytes_per_sec": 0, 00:08:48.078 "w_mbytes_per_sec": 0 00:08:48.078 }, 00:08:48.078 "claimed": false, 00:08:48.078 "zoned": false, 00:08:48.078 "supported_io_types": { 00:08:48.078 "read": true, 00:08:48.078 "write": true, 00:08:48.078 "unmap": true, 00:08:48.078 "flush": true, 00:08:48.078 "reset": true, 00:08:48.078 "nvme_admin": false, 00:08:48.078 "nvme_io": false, 00:08:48.078 "nvme_io_md": false, 00:08:48.078 "write_zeroes": true, 00:08:48.078 "zcopy": false, 00:08:48.078 "get_zone_info": false, 00:08:48.078 "zone_management": false, 00:08:48.078 "zone_append": false, 00:08:48.078 "compare": false, 00:08:48.078 "compare_and_write": false, 00:08:48.078 "abort": false, 00:08:48.078 "seek_hole": false, 00:08:48.078 "seek_data": false, 00:08:48.078 "copy": false, 00:08:48.078 "nvme_iov_md": false 00:08:48.078 }, 00:08:48.078 "memory_domains": [ 00:08:48.078 { 00:08:48.078 "dma_device_id": "system", 00:08:48.078 "dma_device_type": 1 00:08:48.078 }, 00:08:48.078 { 00:08:48.078 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:48.078 "dma_device_type": 2 00:08:48.078 }, 00:08:48.078 { 00:08:48.078 "dma_device_id": "system", 00:08:48.078 "dma_device_type": 1 00:08:48.078 }, 00:08:48.078 { 00:08:48.078 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:48.078 "dma_device_type": 2 00:08:48.078 } 00:08:48.078 ], 00:08:48.078 "driver_specific": { 00:08:48.078 "raid": { 00:08:48.078 "uuid": "8afdb292-a0db-4a39-8de0-cdd529e32bf9", 00:08:48.078 "strip_size_kb": 64, 00:08:48.078 "state": "online", 00:08:48.078 "raid_level": "raid0", 00:08:48.078 "superblock": true, 00:08:48.078 "num_base_bdevs": 2, 00:08:48.078 "num_base_bdevs_discovered": 2, 00:08:48.078 "num_base_bdevs_operational": 2, 00:08:48.078 "base_bdevs_list": [ 00:08:48.078 { 00:08:48.078 "name": "pt1", 00:08:48.078 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:48.078 "is_configured": true, 00:08:48.078 "data_offset": 2048, 00:08:48.078 "data_size": 63488 00:08:48.078 }, 00:08:48.078 { 00:08:48.078 "name": "pt2", 00:08:48.078 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:48.078 "is_configured": true, 00:08:48.078 "data_offset": 2048, 00:08:48.078 "data_size": 63488 00:08:48.078 } 00:08:48.078 ] 00:08:48.078 } 00:08:48.078 } 00:08:48.078 }' 00:08:48.078 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:48.078 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:48.078 pt2' 00:08:48.078 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:48.078 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:48.078 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:48.079 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:48.079 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.079 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.079 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:48.079 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.337 [2024-11-20 15:56:46.370993] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 8afdb292-a0db-4a39-8de0-cdd529e32bf9 '!=' 8afdb292-a0db-4a39-8de0-cdd529e32bf9 ']' 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 59883 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 59883 ']' 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 59883 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 59883 00:08:48.337 killing process with pid 59883 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 59883' 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 59883 00:08:48.337 [2024-11-20 15:56:46.429041] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:48.337 15:56:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 59883 00:08:48.337 [2024-11-20 15:56:46.429121] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:48.337 [2024-11-20 15:56:46.429166] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:48.337 [2024-11-20 15:56:46.429179] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:08:48.337 [2024-11-20 15:56:46.555747] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:49.272 ************************************ 00:08:49.272 END TEST raid_superblock_test 00:08:49.272 ************************************ 00:08:49.272 15:56:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:49.272 00:08:49.272 real 0m3.394s 00:08:49.272 user 0m4.773s 00:08:49.272 sys 0m0.514s 00:08:49.272 15:56:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:49.272 15:56:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.272 15:56:47 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 2 read 00:08:49.272 15:56:47 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:49.272 15:56:47 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:49.272 15:56:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:49.272 ************************************ 00:08:49.272 START TEST raid_read_error_test 00:08:49.272 ************************************ 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 2 read 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:49.272 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.vjTtgLl0Ct 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=60078 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 60078 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 60078 ']' 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.272 15:56:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:49.272 [2024-11-20 15:56:47.372772] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:49.272 [2024-11-20 15:56:47.372907] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid60078 ] 00:08:49.530 [2024-11-20 15:56:47.534941] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:49.530 [2024-11-20 15:56:47.634720] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:49.530 [2024-11-20 15:56:47.770910] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:49.530 [2024-11-20 15:56:47.770951] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:50.097 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:50.097 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:50.097 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:50.097 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:50.097 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.097 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.097 BaseBdev1_malloc 00:08:50.097 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.097 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:50.097 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.097 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.097 true 00:08:50.098 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.098 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:50.098 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.098 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.098 [2024-11-20 15:56:48.304598] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:50.098 [2024-11-20 15:56:48.304803] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:50.098 [2024-11-20 15:56:48.304830] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:50.098 [2024-11-20 15:56:48.304842] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:50.098 [2024-11-20 15:56:48.307153] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:50.098 BaseBdev1 00:08:50.098 [2024-11-20 15:56:48.307308] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:50.098 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.098 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:50.098 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:50.098 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.098 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.098 BaseBdev2_malloc 00:08:50.098 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.098 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:50.098 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.098 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.098 true 00:08:50.098 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.358 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:50.358 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.358 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.358 [2024-11-20 15:56:48.349015] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:50.358 [2024-11-20 15:56:48.349184] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:50.359 [2024-11-20 15:56:48.349224] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:50.359 [2024-11-20 15:56:48.349288] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:50.359 [2024-11-20 15:56:48.351462] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:50.359 [2024-11-20 15:56:48.351581] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:50.359 BaseBdev2 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.359 [2024-11-20 15:56:48.357075] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:50.359 [2024-11-20 15:56:48.359030] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:50.359 [2024-11-20 15:56:48.359298] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:08:50.359 [2024-11-20 15:56:48.359369] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:50.359 [2024-11-20 15:56:48.359649] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:08:50.359 [2024-11-20 15:56:48.359875] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:08:50.359 [2024-11-20 15:56:48.359891] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:08:50.359 [2024-11-20 15:56:48.360046] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:50.359 "name": "raid_bdev1", 00:08:50.359 "uuid": "5eb0396f-378b-4810-ad94-bba8f85930d2", 00:08:50.359 "strip_size_kb": 64, 00:08:50.359 "state": "online", 00:08:50.359 "raid_level": "raid0", 00:08:50.359 "superblock": true, 00:08:50.359 "num_base_bdevs": 2, 00:08:50.359 "num_base_bdevs_discovered": 2, 00:08:50.359 "num_base_bdevs_operational": 2, 00:08:50.359 "base_bdevs_list": [ 00:08:50.359 { 00:08:50.359 "name": "BaseBdev1", 00:08:50.359 "uuid": "56d91e92-45f0-5aaa-9a70-97a8bc48db94", 00:08:50.359 "is_configured": true, 00:08:50.359 "data_offset": 2048, 00:08:50.359 "data_size": 63488 00:08:50.359 }, 00:08:50.359 { 00:08:50.359 "name": "BaseBdev2", 00:08:50.359 "uuid": "0756e665-3610-59b2-b6d9-35721e0a92a6", 00:08:50.359 "is_configured": true, 00:08:50.359 "data_offset": 2048, 00:08:50.359 "data_size": 63488 00:08:50.359 } 00:08:50.359 ] 00:08:50.359 }' 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:50.359 15:56:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.617 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:50.617 15:56:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:50.617 [2024-11-20 15:56:48.746100] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:51.551 "name": "raid_bdev1", 00:08:51.551 "uuid": "5eb0396f-378b-4810-ad94-bba8f85930d2", 00:08:51.551 "strip_size_kb": 64, 00:08:51.551 "state": "online", 00:08:51.551 "raid_level": "raid0", 00:08:51.551 "superblock": true, 00:08:51.551 "num_base_bdevs": 2, 00:08:51.551 "num_base_bdevs_discovered": 2, 00:08:51.551 "num_base_bdevs_operational": 2, 00:08:51.551 "base_bdevs_list": [ 00:08:51.551 { 00:08:51.551 "name": "BaseBdev1", 00:08:51.551 "uuid": "56d91e92-45f0-5aaa-9a70-97a8bc48db94", 00:08:51.551 "is_configured": true, 00:08:51.551 "data_offset": 2048, 00:08:51.551 "data_size": 63488 00:08:51.551 }, 00:08:51.551 { 00:08:51.551 "name": "BaseBdev2", 00:08:51.551 "uuid": "0756e665-3610-59b2-b6d9-35721e0a92a6", 00:08:51.551 "is_configured": true, 00:08:51.551 "data_offset": 2048, 00:08:51.551 "data_size": 63488 00:08:51.551 } 00:08:51.551 ] 00:08:51.551 }' 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:51.551 15:56:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.810 15:56:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:51.810 15:56:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.810 15:56:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.810 [2024-11-20 15:56:50.016266] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:51.810 [2024-11-20 15:56:50.016421] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:51.810 [2024-11-20 15:56:50.019580] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:51.810 { 00:08:51.810 "results": [ 00:08:51.810 { 00:08:51.810 "job": "raid_bdev1", 00:08:51.810 "core_mask": "0x1", 00:08:51.810 "workload": "randrw", 00:08:51.810 "percentage": 50, 00:08:51.810 "status": "finished", 00:08:51.810 "queue_depth": 1, 00:08:51.810 "io_size": 131072, 00:08:51.810 "runtime": 1.268488, 00:08:51.810 "iops": 14633.16956880948, 00:08:51.810 "mibps": 1829.146196101185, 00:08:51.810 "io_failed": 1, 00:08:51.810 "io_timeout": 0, 00:08:51.810 "avg_latency_us": 93.18656616345997, 00:08:51.810 "min_latency_us": 33.673846153846156, 00:08:51.810 "max_latency_us": 1688.8123076923077 00:08:51.810 } 00:08:51.810 ], 00:08:51.810 "core_count": 1 00:08:51.810 } 00:08:51.810 [2024-11-20 15:56:50.019734] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:51.810 [2024-11-20 15:56:50.019774] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:51.810 [2024-11-20 15:56:50.019787] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:08:51.810 15:56:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.810 15:56:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 60078 00:08:51.810 15:56:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 60078 ']' 00:08:51.810 15:56:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 60078 00:08:51.810 15:56:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:08:51.810 15:56:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:51.810 15:56:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 60078 00:08:51.810 15:56:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:51.810 15:56:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:51.810 15:56:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 60078' 00:08:51.810 killing process with pid 60078 00:08:51.810 15:56:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 60078 00:08:51.810 [2024-11-20 15:56:50.050905] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:51.810 15:56:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 60078 00:08:52.068 [2024-11-20 15:56:50.135308] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:52.633 15:56:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.vjTtgLl0Ct 00:08:52.633 15:56:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:52.633 15:56:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:52.633 15:56:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.79 00:08:52.633 15:56:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:52.633 15:56:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:52.633 15:56:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:52.633 15:56:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.79 != \0\.\0\0 ]] 00:08:52.633 00:08:52.633 real 0m3.561s 00:08:52.633 user 0m4.326s 00:08:52.633 sys 0m0.370s 00:08:52.633 ************************************ 00:08:52.633 END TEST raid_read_error_test 00:08:52.633 ************************************ 00:08:52.633 15:56:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:52.633 15:56:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.891 15:56:50 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 2 write 00:08:52.891 15:56:50 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:52.891 15:56:50 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:52.891 15:56:50 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:52.891 ************************************ 00:08:52.891 START TEST raid_write_error_test 00:08:52.891 ************************************ 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 2 write 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.6iyZJLR22n 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=60213 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 60213 00:08:52.891 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 60213 ']' 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.891 15:56:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:52.891 [2024-11-20 15:56:50.965595] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:52.891 [2024-11-20 15:56:50.965710] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid60213 ] 00:08:52.891 [2024-11-20 15:56:51.116173] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:53.149 [2024-11-20 15:56:51.214964] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:53.149 [2024-11-20 15:56:51.328096] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:53.149 [2024-11-20 15:56:51.328145] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:53.749 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:53.749 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:53.749 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:53.749 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:53.749 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.749 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.749 BaseBdev1_malloc 00:08:53.749 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.749 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:53.749 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.749 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.749 true 00:08:53.749 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.749 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:53.749 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.749 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.749 [2024-11-20 15:56:51.862862] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:53.749 [2024-11-20 15:56:51.862914] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:53.749 [2024-11-20 15:56:51.862931] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:53.749 [2024-11-20 15:56:51.862939] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:53.749 [2024-11-20 15:56:51.864756] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:53.749 [2024-11-20 15:56:51.864894] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:53.749 BaseBdev1 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.750 BaseBdev2_malloc 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.750 true 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.750 [2024-11-20 15:56:51.902405] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:53.750 [2024-11-20 15:56:51.902475] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:53.750 [2024-11-20 15:56:51.902492] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:53.750 [2024-11-20 15:56:51.902502] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:53.750 [2024-11-20 15:56:51.904286] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:53.750 [2024-11-20 15:56:51.904427] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:53.750 BaseBdev2 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.750 [2024-11-20 15:56:51.910459] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:53.750 [2024-11-20 15:56:51.912096] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:53.750 [2024-11-20 15:56:51.912315] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:08:53.750 [2024-11-20 15:56:51.912384] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:53.750 [2024-11-20 15:56:51.912618] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:08:53.750 [2024-11-20 15:56:51.912773] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:08:53.750 [2024-11-20 15:56:51.912798] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:08:53.750 [2024-11-20 15:56:51.912985] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:53.750 "name": "raid_bdev1", 00:08:53.750 "uuid": "a3ab8236-43dc-4a35-8317-a7275ace0001", 00:08:53.750 "strip_size_kb": 64, 00:08:53.750 "state": "online", 00:08:53.750 "raid_level": "raid0", 00:08:53.750 "superblock": true, 00:08:53.750 "num_base_bdevs": 2, 00:08:53.750 "num_base_bdevs_discovered": 2, 00:08:53.750 "num_base_bdevs_operational": 2, 00:08:53.750 "base_bdevs_list": [ 00:08:53.750 { 00:08:53.750 "name": "BaseBdev1", 00:08:53.750 "uuid": "7715868f-51a0-5128-83e2-a5d26f1ecbcf", 00:08:53.750 "is_configured": true, 00:08:53.750 "data_offset": 2048, 00:08:53.750 "data_size": 63488 00:08:53.750 }, 00:08:53.750 { 00:08:53.750 "name": "BaseBdev2", 00:08:53.750 "uuid": "c00436bd-3dcf-51d7-9cf3-328e9639d8ec", 00:08:53.750 "is_configured": true, 00:08:53.750 "data_offset": 2048, 00:08:53.750 "data_size": 63488 00:08:53.750 } 00:08:53.750 ] 00:08:53.750 }' 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:53.750 15:56:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.026 15:56:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:54.026 15:56:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:54.283 [2024-11-20 15:56:52.319283] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:08:55.217 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:55.218 "name": "raid_bdev1", 00:08:55.218 "uuid": "a3ab8236-43dc-4a35-8317-a7275ace0001", 00:08:55.218 "strip_size_kb": 64, 00:08:55.218 "state": "online", 00:08:55.218 "raid_level": "raid0", 00:08:55.218 "superblock": true, 00:08:55.218 "num_base_bdevs": 2, 00:08:55.218 "num_base_bdevs_discovered": 2, 00:08:55.218 "num_base_bdevs_operational": 2, 00:08:55.218 "base_bdevs_list": [ 00:08:55.218 { 00:08:55.218 "name": "BaseBdev1", 00:08:55.218 "uuid": "7715868f-51a0-5128-83e2-a5d26f1ecbcf", 00:08:55.218 "is_configured": true, 00:08:55.218 "data_offset": 2048, 00:08:55.218 "data_size": 63488 00:08:55.218 }, 00:08:55.218 { 00:08:55.218 "name": "BaseBdev2", 00:08:55.218 "uuid": "c00436bd-3dcf-51d7-9cf3-328e9639d8ec", 00:08:55.218 "is_configured": true, 00:08:55.218 "data_offset": 2048, 00:08:55.218 "data_size": 63488 00:08:55.218 } 00:08:55.218 ] 00:08:55.218 }' 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:55.218 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.476 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:55.476 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.476 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.476 [2024-11-20 15:56:53.550626] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:55.476 [2024-11-20 15:56:53.550658] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:55.476 [2024-11-20 15:56:53.553150] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:55.476 [2024-11-20 15:56:53.553295] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:55.476 [2024-11-20 15:56:53.553331] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:55.476 [2024-11-20 15:56:53.553341] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:08:55.476 { 00:08:55.476 "results": [ 00:08:55.476 { 00:08:55.476 "job": "raid_bdev1", 00:08:55.476 "core_mask": "0x1", 00:08:55.476 "workload": "randrw", 00:08:55.476 "percentage": 50, 00:08:55.476 "status": "finished", 00:08:55.476 "queue_depth": 1, 00:08:55.476 "io_size": 131072, 00:08:55.476 "runtime": 1.229938, 00:08:55.476 "iops": 18008.22480482756, 00:08:55.476 "mibps": 2251.028100603445, 00:08:55.476 "io_failed": 1, 00:08:55.476 "io_timeout": 0, 00:08:55.476 "avg_latency_us": 75.84687897204375, 00:08:55.476 "min_latency_us": 25.796923076923076, 00:08:55.476 "max_latency_us": 1304.4184615384615 00:08:55.476 } 00:08:55.476 ], 00:08:55.476 "core_count": 1 00:08:55.476 } 00:08:55.476 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.476 15:56:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 60213 00:08:55.476 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 60213 ']' 00:08:55.476 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 60213 00:08:55.476 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:08:55.476 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:55.476 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 60213 00:08:55.476 killing process with pid 60213 00:08:55.476 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:55.476 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:55.476 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 60213' 00:08:55.476 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 60213 00:08:55.476 [2024-11-20 15:56:53.581493] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:55.476 15:56:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 60213 00:08:55.476 [2024-11-20 15:56:53.648020] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:56.043 15:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.6iyZJLR22n 00:08:56.043 15:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:56.043 15:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:56.043 ************************************ 00:08:56.043 END TEST raid_write_error_test 00:08:56.043 ************************************ 00:08:56.043 15:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.81 00:08:56.043 15:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:56.043 15:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:56.043 15:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:56.043 15:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.81 != \0\.\0\0 ]] 00:08:56.043 00:08:56.043 real 0m3.360s 00:08:56.043 user 0m4.075s 00:08:56.043 sys 0m0.356s 00:08:56.043 15:56:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:56.043 15:56:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.043 15:56:54 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:56.043 15:56:54 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 2 false 00:08:56.043 15:56:54 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:56.043 15:56:54 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:56.043 15:56:54 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:56.301 ************************************ 00:08:56.301 START TEST raid_state_function_test 00:08:56.301 ************************************ 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 2 false 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:56.301 Process raid pid: 60345 00:08:56.301 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=60345 00:08:56.301 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 60345' 00:08:56.302 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 60345 00:08:56.302 15:56:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 60345 ']' 00:08:56.302 15:56:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:56.302 15:56:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:56.302 15:56:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:56.302 15:56:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:56.302 15:56:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:56.302 15:56:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.302 [2024-11-20 15:56:54.356349] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:56.302 [2024-11-20 15:56:54.356454] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:56.302 [2024-11-20 15:56:54.504080] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:56.559 [2024-11-20 15:56:54.588922] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:56.559 [2024-11-20 15:56:54.700290] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:56.559 [2024-11-20 15:56:54.700326] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.124 [2024-11-20 15:56:55.220079] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:57.124 [2024-11-20 15:56:55.220127] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:57.124 [2024-11-20 15:56:55.220136] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:57.124 [2024-11-20 15:56:55.220145] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:57.124 "name": "Existed_Raid", 00:08:57.124 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.124 "strip_size_kb": 64, 00:08:57.124 "state": "configuring", 00:08:57.124 "raid_level": "concat", 00:08:57.124 "superblock": false, 00:08:57.124 "num_base_bdevs": 2, 00:08:57.124 "num_base_bdevs_discovered": 0, 00:08:57.124 "num_base_bdevs_operational": 2, 00:08:57.124 "base_bdevs_list": [ 00:08:57.124 { 00:08:57.124 "name": "BaseBdev1", 00:08:57.124 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.124 "is_configured": false, 00:08:57.124 "data_offset": 0, 00:08:57.124 "data_size": 0 00:08:57.124 }, 00:08:57.124 { 00:08:57.124 "name": "BaseBdev2", 00:08:57.124 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.124 "is_configured": false, 00:08:57.124 "data_offset": 0, 00:08:57.124 "data_size": 0 00:08:57.124 } 00:08:57.124 ] 00:08:57.124 }' 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:57.124 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.383 [2024-11-20 15:56:55.516106] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:57.383 [2024-11-20 15:56:55.516138] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.383 [2024-11-20 15:56:55.524104] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:57.383 [2024-11-20 15:56:55.524144] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:57.383 [2024-11-20 15:56:55.524151] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:57.383 [2024-11-20 15:56:55.524160] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.383 [2024-11-20 15:56:55.552246] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:57.383 BaseBdev1 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.383 [ 00:08:57.383 { 00:08:57.383 "name": "BaseBdev1", 00:08:57.383 "aliases": [ 00:08:57.383 "37e5a8f1-15f4-4259-aba5-b2923da7846a" 00:08:57.383 ], 00:08:57.383 "product_name": "Malloc disk", 00:08:57.383 "block_size": 512, 00:08:57.383 "num_blocks": 65536, 00:08:57.383 "uuid": "37e5a8f1-15f4-4259-aba5-b2923da7846a", 00:08:57.383 "assigned_rate_limits": { 00:08:57.383 "rw_ios_per_sec": 0, 00:08:57.383 "rw_mbytes_per_sec": 0, 00:08:57.383 "r_mbytes_per_sec": 0, 00:08:57.383 "w_mbytes_per_sec": 0 00:08:57.383 }, 00:08:57.383 "claimed": true, 00:08:57.383 "claim_type": "exclusive_write", 00:08:57.383 "zoned": false, 00:08:57.383 "supported_io_types": { 00:08:57.383 "read": true, 00:08:57.383 "write": true, 00:08:57.383 "unmap": true, 00:08:57.383 "flush": true, 00:08:57.383 "reset": true, 00:08:57.383 "nvme_admin": false, 00:08:57.383 "nvme_io": false, 00:08:57.383 "nvme_io_md": false, 00:08:57.383 "write_zeroes": true, 00:08:57.383 "zcopy": true, 00:08:57.383 "get_zone_info": false, 00:08:57.383 "zone_management": false, 00:08:57.383 "zone_append": false, 00:08:57.383 "compare": false, 00:08:57.383 "compare_and_write": false, 00:08:57.383 "abort": true, 00:08:57.383 "seek_hole": false, 00:08:57.383 "seek_data": false, 00:08:57.383 "copy": true, 00:08:57.383 "nvme_iov_md": false 00:08:57.383 }, 00:08:57.383 "memory_domains": [ 00:08:57.383 { 00:08:57.383 "dma_device_id": "system", 00:08:57.383 "dma_device_type": 1 00:08:57.383 }, 00:08:57.383 { 00:08:57.383 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:57.383 "dma_device_type": 2 00:08:57.383 } 00:08:57.383 ], 00:08:57.383 "driver_specific": {} 00:08:57.383 } 00:08:57.383 ] 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:57.383 "name": "Existed_Raid", 00:08:57.383 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.383 "strip_size_kb": 64, 00:08:57.383 "state": "configuring", 00:08:57.383 "raid_level": "concat", 00:08:57.383 "superblock": false, 00:08:57.383 "num_base_bdevs": 2, 00:08:57.383 "num_base_bdevs_discovered": 1, 00:08:57.383 "num_base_bdevs_operational": 2, 00:08:57.383 "base_bdevs_list": [ 00:08:57.383 { 00:08:57.383 "name": "BaseBdev1", 00:08:57.383 "uuid": "37e5a8f1-15f4-4259-aba5-b2923da7846a", 00:08:57.383 "is_configured": true, 00:08:57.383 "data_offset": 0, 00:08:57.383 "data_size": 65536 00:08:57.383 }, 00:08:57.383 { 00:08:57.383 "name": "BaseBdev2", 00:08:57.383 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.383 "is_configured": false, 00:08:57.383 "data_offset": 0, 00:08:57.383 "data_size": 0 00:08:57.383 } 00:08:57.383 ] 00:08:57.383 }' 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:57.383 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.696 [2024-11-20 15:56:55.860340] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:57.696 [2024-11-20 15:56:55.860383] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.696 [2024-11-20 15:56:55.868380] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:57.696 [2024-11-20 15:56:55.869930] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:57.696 [2024-11-20 15:56:55.869965] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:57.696 "name": "Existed_Raid", 00:08:57.696 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.696 "strip_size_kb": 64, 00:08:57.696 "state": "configuring", 00:08:57.696 "raid_level": "concat", 00:08:57.696 "superblock": false, 00:08:57.696 "num_base_bdevs": 2, 00:08:57.696 "num_base_bdevs_discovered": 1, 00:08:57.696 "num_base_bdevs_operational": 2, 00:08:57.696 "base_bdevs_list": [ 00:08:57.696 { 00:08:57.696 "name": "BaseBdev1", 00:08:57.696 "uuid": "37e5a8f1-15f4-4259-aba5-b2923da7846a", 00:08:57.696 "is_configured": true, 00:08:57.696 "data_offset": 0, 00:08:57.696 "data_size": 65536 00:08:57.696 }, 00:08:57.696 { 00:08:57.696 "name": "BaseBdev2", 00:08:57.696 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.696 "is_configured": false, 00:08:57.696 "data_offset": 0, 00:08:57.696 "data_size": 0 00:08:57.696 } 00:08:57.696 ] 00:08:57.696 }' 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:57.696 15:56:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.969 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:57.969 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.969 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.969 [2024-11-20 15:56:56.214861] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:57.969 [2024-11-20 15:56:56.215066] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:08:57.969 [2024-11-20 15:56:56.215079] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:08:57.969 [2024-11-20 15:56:56.215310] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:57.969 [2024-11-20 15:56:56.215457] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:08:57.969 [2024-11-20 15:56:56.215466] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:08:57.969 [2024-11-20 15:56:56.215698] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:57.969 BaseBdev2 00:08:57.969 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.969 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:57.969 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:57.969 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:57.969 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:57.969 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:57.969 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:57.970 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:57.970 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.970 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.227 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.227 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:58.227 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.227 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.227 [ 00:08:58.227 { 00:08:58.227 "name": "BaseBdev2", 00:08:58.227 "aliases": [ 00:08:58.227 "a932246e-9152-47d6-8aae-a1dc1f4c6ed4" 00:08:58.227 ], 00:08:58.227 "product_name": "Malloc disk", 00:08:58.227 "block_size": 512, 00:08:58.227 "num_blocks": 65536, 00:08:58.227 "uuid": "a932246e-9152-47d6-8aae-a1dc1f4c6ed4", 00:08:58.227 "assigned_rate_limits": { 00:08:58.227 "rw_ios_per_sec": 0, 00:08:58.227 "rw_mbytes_per_sec": 0, 00:08:58.227 "r_mbytes_per_sec": 0, 00:08:58.227 "w_mbytes_per_sec": 0 00:08:58.227 }, 00:08:58.227 "claimed": true, 00:08:58.227 "claim_type": "exclusive_write", 00:08:58.227 "zoned": false, 00:08:58.227 "supported_io_types": { 00:08:58.227 "read": true, 00:08:58.227 "write": true, 00:08:58.227 "unmap": true, 00:08:58.227 "flush": true, 00:08:58.227 "reset": true, 00:08:58.227 "nvme_admin": false, 00:08:58.227 "nvme_io": false, 00:08:58.227 "nvme_io_md": false, 00:08:58.227 "write_zeroes": true, 00:08:58.227 "zcopy": true, 00:08:58.227 "get_zone_info": false, 00:08:58.227 "zone_management": false, 00:08:58.227 "zone_append": false, 00:08:58.227 "compare": false, 00:08:58.227 "compare_and_write": false, 00:08:58.227 "abort": true, 00:08:58.227 "seek_hole": false, 00:08:58.227 "seek_data": false, 00:08:58.227 "copy": true, 00:08:58.227 "nvme_iov_md": false 00:08:58.227 }, 00:08:58.227 "memory_domains": [ 00:08:58.227 { 00:08:58.227 "dma_device_id": "system", 00:08:58.227 "dma_device_type": 1 00:08:58.227 }, 00:08:58.227 { 00:08:58.227 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.227 "dma_device_type": 2 00:08:58.227 } 00:08:58.227 ], 00:08:58.227 "driver_specific": {} 00:08:58.227 } 00:08:58.227 ] 00:08:58.227 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.227 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:58.227 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:58.227 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:58.227 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:08:58.227 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:58.227 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:58.227 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:58.227 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:58.227 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:58.227 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:58.227 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:58.228 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:58.228 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:58.228 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:58.228 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.228 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.228 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.228 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.228 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:58.228 "name": "Existed_Raid", 00:08:58.228 "uuid": "f0461bd7-0d9e-48d6-9284-c5172bf0f3de", 00:08:58.228 "strip_size_kb": 64, 00:08:58.228 "state": "online", 00:08:58.228 "raid_level": "concat", 00:08:58.228 "superblock": false, 00:08:58.228 "num_base_bdevs": 2, 00:08:58.228 "num_base_bdevs_discovered": 2, 00:08:58.228 "num_base_bdevs_operational": 2, 00:08:58.228 "base_bdevs_list": [ 00:08:58.228 { 00:08:58.228 "name": "BaseBdev1", 00:08:58.228 "uuid": "37e5a8f1-15f4-4259-aba5-b2923da7846a", 00:08:58.228 "is_configured": true, 00:08:58.228 "data_offset": 0, 00:08:58.228 "data_size": 65536 00:08:58.228 }, 00:08:58.228 { 00:08:58.228 "name": "BaseBdev2", 00:08:58.228 "uuid": "a932246e-9152-47d6-8aae-a1dc1f4c6ed4", 00:08:58.228 "is_configured": true, 00:08:58.228 "data_offset": 0, 00:08:58.228 "data_size": 65536 00:08:58.228 } 00:08:58.228 ] 00:08:58.228 }' 00:08:58.228 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:58.228 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.486 [2024-11-20 15:56:56.555195] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:58.486 "name": "Existed_Raid", 00:08:58.486 "aliases": [ 00:08:58.486 "f0461bd7-0d9e-48d6-9284-c5172bf0f3de" 00:08:58.486 ], 00:08:58.486 "product_name": "Raid Volume", 00:08:58.486 "block_size": 512, 00:08:58.486 "num_blocks": 131072, 00:08:58.486 "uuid": "f0461bd7-0d9e-48d6-9284-c5172bf0f3de", 00:08:58.486 "assigned_rate_limits": { 00:08:58.486 "rw_ios_per_sec": 0, 00:08:58.486 "rw_mbytes_per_sec": 0, 00:08:58.486 "r_mbytes_per_sec": 0, 00:08:58.486 "w_mbytes_per_sec": 0 00:08:58.486 }, 00:08:58.486 "claimed": false, 00:08:58.486 "zoned": false, 00:08:58.486 "supported_io_types": { 00:08:58.486 "read": true, 00:08:58.486 "write": true, 00:08:58.486 "unmap": true, 00:08:58.486 "flush": true, 00:08:58.486 "reset": true, 00:08:58.486 "nvme_admin": false, 00:08:58.486 "nvme_io": false, 00:08:58.486 "nvme_io_md": false, 00:08:58.486 "write_zeroes": true, 00:08:58.486 "zcopy": false, 00:08:58.486 "get_zone_info": false, 00:08:58.486 "zone_management": false, 00:08:58.486 "zone_append": false, 00:08:58.486 "compare": false, 00:08:58.486 "compare_and_write": false, 00:08:58.486 "abort": false, 00:08:58.486 "seek_hole": false, 00:08:58.486 "seek_data": false, 00:08:58.486 "copy": false, 00:08:58.486 "nvme_iov_md": false 00:08:58.486 }, 00:08:58.486 "memory_domains": [ 00:08:58.486 { 00:08:58.486 "dma_device_id": "system", 00:08:58.486 "dma_device_type": 1 00:08:58.486 }, 00:08:58.486 { 00:08:58.486 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.486 "dma_device_type": 2 00:08:58.486 }, 00:08:58.486 { 00:08:58.486 "dma_device_id": "system", 00:08:58.486 "dma_device_type": 1 00:08:58.486 }, 00:08:58.486 { 00:08:58.486 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.486 "dma_device_type": 2 00:08:58.486 } 00:08:58.486 ], 00:08:58.486 "driver_specific": { 00:08:58.486 "raid": { 00:08:58.486 "uuid": "f0461bd7-0d9e-48d6-9284-c5172bf0f3de", 00:08:58.486 "strip_size_kb": 64, 00:08:58.486 "state": "online", 00:08:58.486 "raid_level": "concat", 00:08:58.486 "superblock": false, 00:08:58.486 "num_base_bdevs": 2, 00:08:58.486 "num_base_bdevs_discovered": 2, 00:08:58.486 "num_base_bdevs_operational": 2, 00:08:58.486 "base_bdevs_list": [ 00:08:58.486 { 00:08:58.486 "name": "BaseBdev1", 00:08:58.486 "uuid": "37e5a8f1-15f4-4259-aba5-b2923da7846a", 00:08:58.486 "is_configured": true, 00:08:58.486 "data_offset": 0, 00:08:58.486 "data_size": 65536 00:08:58.486 }, 00:08:58.486 { 00:08:58.486 "name": "BaseBdev2", 00:08:58.486 "uuid": "a932246e-9152-47d6-8aae-a1dc1f4c6ed4", 00:08:58.486 "is_configured": true, 00:08:58.486 "data_offset": 0, 00:08:58.486 "data_size": 65536 00:08:58.486 } 00:08:58.486 ] 00:08:58.486 } 00:08:58.486 } 00:08:58.486 }' 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:58.486 BaseBdev2' 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.486 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.486 [2024-11-20 15:56:56.711027] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:58.486 [2024-11-20 15:56:56.711057] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:58.486 [2024-11-20 15:56:56.711099] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:58.745 "name": "Existed_Raid", 00:08:58.745 "uuid": "f0461bd7-0d9e-48d6-9284-c5172bf0f3de", 00:08:58.745 "strip_size_kb": 64, 00:08:58.745 "state": "offline", 00:08:58.745 "raid_level": "concat", 00:08:58.745 "superblock": false, 00:08:58.745 "num_base_bdevs": 2, 00:08:58.745 "num_base_bdevs_discovered": 1, 00:08:58.745 "num_base_bdevs_operational": 1, 00:08:58.745 "base_bdevs_list": [ 00:08:58.745 { 00:08:58.745 "name": null, 00:08:58.745 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:58.745 "is_configured": false, 00:08:58.745 "data_offset": 0, 00:08:58.745 "data_size": 65536 00:08:58.745 }, 00:08:58.745 { 00:08:58.745 "name": "BaseBdev2", 00:08:58.745 "uuid": "a932246e-9152-47d6-8aae-a1dc1f4c6ed4", 00:08:58.745 "is_configured": true, 00:08:58.745 "data_offset": 0, 00:08:58.745 "data_size": 65536 00:08:58.745 } 00:08:58.745 ] 00:08:58.745 }' 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:58.745 15:56:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.003 [2024-11-20 15:56:57.090381] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:59.003 [2024-11-20 15:56:57.090535] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 60345 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 60345 ']' 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 60345 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 60345 00:08:59.003 killing process with pid 60345 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 60345' 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 60345 00:08:59.003 [2024-11-20 15:56:57.202189] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:59.003 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 60345 00:08:59.003 [2024-11-20 15:56:57.210606] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:59.567 15:56:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:59.567 00:08:59.567 real 0m3.492s 00:08:59.567 user 0m5.097s 00:08:59.567 sys 0m0.551s 00:08:59.567 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:59.567 15:56:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.567 ************************************ 00:08:59.567 END TEST raid_state_function_test 00:08:59.567 ************************************ 00:08:59.824 15:56:57 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 2 true 00:08:59.824 15:56:57 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:59.824 15:56:57 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:59.824 15:56:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:59.824 ************************************ 00:08:59.824 START TEST raid_state_function_test_sb 00:08:59.824 ************************************ 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 2 true 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:59.824 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:59.825 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:59.825 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:08:59.825 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:59.825 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:59.825 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:59.825 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:59.825 Process raid pid: 60580 00:08:59.825 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:59.825 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=60580 00:08:59.825 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 60580' 00:08:59.825 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 60580 00:08:59.825 15:56:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:59.825 15:56:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 60580 ']' 00:08:59.825 15:56:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:59.825 15:56:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:59.825 15:56:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:59.825 15:56:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:59.825 15:56:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.825 [2024-11-20 15:56:57.902527] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:08:59.825 [2024-11-20 15:56:57.902804] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:59.825 [2024-11-20 15:56:58.063569] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:00.082 [2024-11-20 15:56:58.161977] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:00.339 [2024-11-20 15:56:58.340647] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:00.339 [2024-11-20 15:56:58.340808] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:00.596 15:56:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:00.596 15:56:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:09:00.596 15:56:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:09:00.596 15:56:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.596 15:56:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.596 [2024-11-20 15:56:58.754786] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:00.596 [2024-11-20 15:56:58.754844] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:00.596 [2024-11-20 15:56:58.754858] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:00.596 [2024-11-20 15:56:58.754868] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:00.596 15:56:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.596 15:56:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:09:00.596 15:56:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:00.596 15:56:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:00.596 15:56:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:00.597 15:56:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:00.597 15:56:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:00.597 15:56:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:00.597 15:56:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:00.597 15:56:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:00.597 15:56:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:00.597 15:56:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:00.597 15:56:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:00.597 15:56:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.597 15:56:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.597 15:56:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.597 15:56:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:00.597 "name": "Existed_Raid", 00:09:00.597 "uuid": "0d9c037f-bbca-4164-aa0e-af06bb5fea48", 00:09:00.597 "strip_size_kb": 64, 00:09:00.597 "state": "configuring", 00:09:00.597 "raid_level": "concat", 00:09:00.597 "superblock": true, 00:09:00.597 "num_base_bdevs": 2, 00:09:00.597 "num_base_bdevs_discovered": 0, 00:09:00.597 "num_base_bdevs_operational": 2, 00:09:00.597 "base_bdevs_list": [ 00:09:00.597 { 00:09:00.597 "name": "BaseBdev1", 00:09:00.597 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.597 "is_configured": false, 00:09:00.597 "data_offset": 0, 00:09:00.597 "data_size": 0 00:09:00.597 }, 00:09:00.597 { 00:09:00.597 "name": "BaseBdev2", 00:09:00.597 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.597 "is_configured": false, 00:09:00.597 "data_offset": 0, 00:09:00.597 "data_size": 0 00:09:00.597 } 00:09:00.597 ] 00:09:00.597 }' 00:09:00.597 15:56:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:00.597 15:56:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.854 [2024-11-20 15:56:59.058804] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:00.854 [2024-11-20 15:56:59.058841] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.854 [2024-11-20 15:56:59.066813] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:00.854 [2024-11-20 15:56:59.066854] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:00.854 [2024-11-20 15:56:59.066863] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:00.854 [2024-11-20 15:56:59.066874] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.854 [2024-11-20 15:56:59.099022] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:00.854 BaseBdev1 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.854 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.114 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.114 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:01.114 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.114 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.114 [ 00:09:01.114 { 00:09:01.114 "name": "BaseBdev1", 00:09:01.114 "aliases": [ 00:09:01.114 "f14c6127-4084-4120-afd0-fcbf7c547936" 00:09:01.114 ], 00:09:01.114 "product_name": "Malloc disk", 00:09:01.114 "block_size": 512, 00:09:01.114 "num_blocks": 65536, 00:09:01.114 "uuid": "f14c6127-4084-4120-afd0-fcbf7c547936", 00:09:01.114 "assigned_rate_limits": { 00:09:01.114 "rw_ios_per_sec": 0, 00:09:01.114 "rw_mbytes_per_sec": 0, 00:09:01.114 "r_mbytes_per_sec": 0, 00:09:01.114 "w_mbytes_per_sec": 0 00:09:01.114 }, 00:09:01.114 "claimed": true, 00:09:01.114 "claim_type": "exclusive_write", 00:09:01.114 "zoned": false, 00:09:01.114 "supported_io_types": { 00:09:01.114 "read": true, 00:09:01.114 "write": true, 00:09:01.114 "unmap": true, 00:09:01.114 "flush": true, 00:09:01.114 "reset": true, 00:09:01.114 "nvme_admin": false, 00:09:01.114 "nvme_io": false, 00:09:01.114 "nvme_io_md": false, 00:09:01.114 "write_zeroes": true, 00:09:01.114 "zcopy": true, 00:09:01.114 "get_zone_info": false, 00:09:01.114 "zone_management": false, 00:09:01.114 "zone_append": false, 00:09:01.114 "compare": false, 00:09:01.114 "compare_and_write": false, 00:09:01.114 "abort": true, 00:09:01.114 "seek_hole": false, 00:09:01.114 "seek_data": false, 00:09:01.114 "copy": true, 00:09:01.114 "nvme_iov_md": false 00:09:01.114 }, 00:09:01.114 "memory_domains": [ 00:09:01.114 { 00:09:01.114 "dma_device_id": "system", 00:09:01.114 "dma_device_type": 1 00:09:01.114 }, 00:09:01.114 { 00:09:01.114 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:01.114 "dma_device_type": 2 00:09:01.114 } 00:09:01.114 ], 00:09:01.114 "driver_specific": {} 00:09:01.114 } 00:09:01.114 ] 00:09:01.114 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.114 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:01.114 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:09:01.114 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:01.115 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:01.115 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:01.115 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:01.115 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:01.115 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:01.115 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:01.115 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:01.115 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:01.115 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.115 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:01.115 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.115 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.115 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.115 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:01.115 "name": "Existed_Raid", 00:09:01.115 "uuid": "b10df816-702a-4f83-839d-fff82076db8b", 00:09:01.115 "strip_size_kb": 64, 00:09:01.115 "state": "configuring", 00:09:01.115 "raid_level": "concat", 00:09:01.115 "superblock": true, 00:09:01.115 "num_base_bdevs": 2, 00:09:01.115 "num_base_bdevs_discovered": 1, 00:09:01.115 "num_base_bdevs_operational": 2, 00:09:01.115 "base_bdevs_list": [ 00:09:01.115 { 00:09:01.115 "name": "BaseBdev1", 00:09:01.115 "uuid": "f14c6127-4084-4120-afd0-fcbf7c547936", 00:09:01.115 "is_configured": true, 00:09:01.115 "data_offset": 2048, 00:09:01.115 "data_size": 63488 00:09:01.115 }, 00:09:01.115 { 00:09:01.115 "name": "BaseBdev2", 00:09:01.115 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:01.115 "is_configured": false, 00:09:01.115 "data_offset": 0, 00:09:01.115 "data_size": 0 00:09:01.115 } 00:09:01.115 ] 00:09:01.115 }' 00:09:01.115 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:01.115 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.373 [2024-11-20 15:56:59.423196] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:01.373 [2024-11-20 15:56:59.423390] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.373 [2024-11-20 15:56:59.431274] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:01.373 [2024-11-20 15:56:59.433857] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:01.373 [2024-11-20 15:56:59.433909] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.373 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:01.373 "name": "Existed_Raid", 00:09:01.373 "uuid": "c67ba1fc-f756-4e5a-8a4d-c60ad90f5262", 00:09:01.373 "strip_size_kb": 64, 00:09:01.373 "state": "configuring", 00:09:01.373 "raid_level": "concat", 00:09:01.373 "superblock": true, 00:09:01.373 "num_base_bdevs": 2, 00:09:01.374 "num_base_bdevs_discovered": 1, 00:09:01.374 "num_base_bdevs_operational": 2, 00:09:01.374 "base_bdevs_list": [ 00:09:01.374 { 00:09:01.374 "name": "BaseBdev1", 00:09:01.374 "uuid": "f14c6127-4084-4120-afd0-fcbf7c547936", 00:09:01.374 "is_configured": true, 00:09:01.374 "data_offset": 2048, 00:09:01.374 "data_size": 63488 00:09:01.374 }, 00:09:01.374 { 00:09:01.374 "name": "BaseBdev2", 00:09:01.374 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:01.374 "is_configured": false, 00:09:01.374 "data_offset": 0, 00:09:01.374 "data_size": 0 00:09:01.374 } 00:09:01.374 ] 00:09:01.374 }' 00:09:01.374 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:01.374 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.631 [2024-11-20 15:56:59.781965] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:01.631 [2024-11-20 15:56:59.782186] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:01.631 [2024-11-20 15:56:59.782201] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:09:01.631 BaseBdev2 00:09:01.631 [2024-11-20 15:56:59.782457] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:09:01.631 [2024-11-20 15:56:59.782594] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:01.631 [2024-11-20 15:56:59.782606] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:09:01.631 [2024-11-20 15:56:59.782741] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.631 [ 00:09:01.631 { 00:09:01.631 "name": "BaseBdev2", 00:09:01.631 "aliases": [ 00:09:01.631 "7298fa02-a931-4009-ad53-a93824e2aa5a" 00:09:01.631 ], 00:09:01.631 "product_name": "Malloc disk", 00:09:01.631 "block_size": 512, 00:09:01.631 "num_blocks": 65536, 00:09:01.631 "uuid": "7298fa02-a931-4009-ad53-a93824e2aa5a", 00:09:01.631 "assigned_rate_limits": { 00:09:01.631 "rw_ios_per_sec": 0, 00:09:01.631 "rw_mbytes_per_sec": 0, 00:09:01.631 "r_mbytes_per_sec": 0, 00:09:01.631 "w_mbytes_per_sec": 0 00:09:01.631 }, 00:09:01.631 "claimed": true, 00:09:01.631 "claim_type": "exclusive_write", 00:09:01.631 "zoned": false, 00:09:01.631 "supported_io_types": { 00:09:01.631 "read": true, 00:09:01.631 "write": true, 00:09:01.631 "unmap": true, 00:09:01.631 "flush": true, 00:09:01.631 "reset": true, 00:09:01.631 "nvme_admin": false, 00:09:01.631 "nvme_io": false, 00:09:01.631 "nvme_io_md": false, 00:09:01.631 "write_zeroes": true, 00:09:01.631 "zcopy": true, 00:09:01.631 "get_zone_info": false, 00:09:01.631 "zone_management": false, 00:09:01.631 "zone_append": false, 00:09:01.631 "compare": false, 00:09:01.631 "compare_and_write": false, 00:09:01.631 "abort": true, 00:09:01.631 "seek_hole": false, 00:09:01.631 "seek_data": false, 00:09:01.631 "copy": true, 00:09:01.631 "nvme_iov_md": false 00:09:01.631 }, 00:09:01.631 "memory_domains": [ 00:09:01.631 { 00:09:01.631 "dma_device_id": "system", 00:09:01.631 "dma_device_type": 1 00:09:01.631 }, 00:09:01.631 { 00:09:01.631 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:01.631 "dma_device_type": 2 00:09:01.631 } 00:09:01.631 ], 00:09:01.631 "driver_specific": {} 00:09:01.631 } 00:09:01.631 ] 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:01.631 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:01.632 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:01.632 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:01.632 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:01.632 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:01.632 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:01.632 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:01.632 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.632 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:01.632 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.632 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.632 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.632 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:01.632 "name": "Existed_Raid", 00:09:01.632 "uuid": "c67ba1fc-f756-4e5a-8a4d-c60ad90f5262", 00:09:01.632 "strip_size_kb": 64, 00:09:01.632 "state": "online", 00:09:01.632 "raid_level": "concat", 00:09:01.632 "superblock": true, 00:09:01.632 "num_base_bdevs": 2, 00:09:01.632 "num_base_bdevs_discovered": 2, 00:09:01.632 "num_base_bdevs_operational": 2, 00:09:01.632 "base_bdevs_list": [ 00:09:01.632 { 00:09:01.632 "name": "BaseBdev1", 00:09:01.632 "uuid": "f14c6127-4084-4120-afd0-fcbf7c547936", 00:09:01.632 "is_configured": true, 00:09:01.632 "data_offset": 2048, 00:09:01.632 "data_size": 63488 00:09:01.632 }, 00:09:01.632 { 00:09:01.632 "name": "BaseBdev2", 00:09:01.632 "uuid": "7298fa02-a931-4009-ad53-a93824e2aa5a", 00:09:01.632 "is_configured": true, 00:09:01.632 "data_offset": 2048, 00:09:01.632 "data_size": 63488 00:09:01.632 } 00:09:01.632 ] 00:09:01.632 }' 00:09:01.632 15:56:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:01.632 15:56:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.889 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:01.889 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:01.889 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:01.889 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:01.889 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:01.889 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:01.889 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:01.889 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:01.889 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.889 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.889 [2024-11-20 15:57:00.082352] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:01.889 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.889 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:01.889 "name": "Existed_Raid", 00:09:01.889 "aliases": [ 00:09:01.889 "c67ba1fc-f756-4e5a-8a4d-c60ad90f5262" 00:09:01.889 ], 00:09:01.889 "product_name": "Raid Volume", 00:09:01.889 "block_size": 512, 00:09:01.889 "num_blocks": 126976, 00:09:01.889 "uuid": "c67ba1fc-f756-4e5a-8a4d-c60ad90f5262", 00:09:01.889 "assigned_rate_limits": { 00:09:01.889 "rw_ios_per_sec": 0, 00:09:01.889 "rw_mbytes_per_sec": 0, 00:09:01.889 "r_mbytes_per_sec": 0, 00:09:01.889 "w_mbytes_per_sec": 0 00:09:01.889 }, 00:09:01.889 "claimed": false, 00:09:01.889 "zoned": false, 00:09:01.889 "supported_io_types": { 00:09:01.889 "read": true, 00:09:01.889 "write": true, 00:09:01.889 "unmap": true, 00:09:01.889 "flush": true, 00:09:01.889 "reset": true, 00:09:01.889 "nvme_admin": false, 00:09:01.889 "nvme_io": false, 00:09:01.889 "nvme_io_md": false, 00:09:01.889 "write_zeroes": true, 00:09:01.889 "zcopy": false, 00:09:01.889 "get_zone_info": false, 00:09:01.889 "zone_management": false, 00:09:01.889 "zone_append": false, 00:09:01.889 "compare": false, 00:09:01.889 "compare_and_write": false, 00:09:01.889 "abort": false, 00:09:01.889 "seek_hole": false, 00:09:01.889 "seek_data": false, 00:09:01.889 "copy": false, 00:09:01.889 "nvme_iov_md": false 00:09:01.889 }, 00:09:01.889 "memory_domains": [ 00:09:01.889 { 00:09:01.889 "dma_device_id": "system", 00:09:01.889 "dma_device_type": 1 00:09:01.889 }, 00:09:01.889 { 00:09:01.889 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:01.889 "dma_device_type": 2 00:09:01.889 }, 00:09:01.889 { 00:09:01.889 "dma_device_id": "system", 00:09:01.889 "dma_device_type": 1 00:09:01.889 }, 00:09:01.889 { 00:09:01.889 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:01.889 "dma_device_type": 2 00:09:01.889 } 00:09:01.889 ], 00:09:01.889 "driver_specific": { 00:09:01.890 "raid": { 00:09:01.890 "uuid": "c67ba1fc-f756-4e5a-8a4d-c60ad90f5262", 00:09:01.890 "strip_size_kb": 64, 00:09:01.890 "state": "online", 00:09:01.890 "raid_level": "concat", 00:09:01.890 "superblock": true, 00:09:01.890 "num_base_bdevs": 2, 00:09:01.890 "num_base_bdevs_discovered": 2, 00:09:01.890 "num_base_bdevs_operational": 2, 00:09:01.890 "base_bdevs_list": [ 00:09:01.890 { 00:09:01.890 "name": "BaseBdev1", 00:09:01.890 "uuid": "f14c6127-4084-4120-afd0-fcbf7c547936", 00:09:01.890 "is_configured": true, 00:09:01.890 "data_offset": 2048, 00:09:01.890 "data_size": 63488 00:09:01.890 }, 00:09:01.890 { 00:09:01.890 "name": "BaseBdev2", 00:09:01.890 "uuid": "7298fa02-a931-4009-ad53-a93824e2aa5a", 00:09:01.890 "is_configured": true, 00:09:01.890 "data_offset": 2048, 00:09:01.890 "data_size": 63488 00:09:01.890 } 00:09:01.890 ] 00:09:01.890 } 00:09:01.890 } 00:09:01.890 }' 00:09:01.890 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:01.890 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:01.890 BaseBdev2' 00:09:01.890 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.148 [2024-11-20 15:57:00.230152] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:02.148 [2024-11-20 15:57:00.230184] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:02.148 [2024-11-20 15:57:00.230229] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:02.148 "name": "Existed_Raid", 00:09:02.148 "uuid": "c67ba1fc-f756-4e5a-8a4d-c60ad90f5262", 00:09:02.148 "strip_size_kb": 64, 00:09:02.148 "state": "offline", 00:09:02.148 "raid_level": "concat", 00:09:02.148 "superblock": true, 00:09:02.148 "num_base_bdevs": 2, 00:09:02.148 "num_base_bdevs_discovered": 1, 00:09:02.148 "num_base_bdevs_operational": 1, 00:09:02.148 "base_bdevs_list": [ 00:09:02.148 { 00:09:02.148 "name": null, 00:09:02.148 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:02.148 "is_configured": false, 00:09:02.148 "data_offset": 0, 00:09:02.148 "data_size": 63488 00:09:02.148 }, 00:09:02.148 { 00:09:02.148 "name": "BaseBdev2", 00:09:02.148 "uuid": "7298fa02-a931-4009-ad53-a93824e2aa5a", 00:09:02.148 "is_configured": true, 00:09:02.148 "data_offset": 2048, 00:09:02.148 "data_size": 63488 00:09:02.148 } 00:09:02.148 ] 00:09:02.148 }' 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:02.148 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.407 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:02.407 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:02.407 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:02.407 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.407 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.407 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.407 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.407 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:02.407 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:02.407 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:02.407 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.407 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.407 [2024-11-20 15:57:00.605327] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:02.407 [2024-11-20 15:57:00.605507] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 60580 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 60580 ']' 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 60580 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 60580 00:09:02.665 killing process with pid 60580 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 60580' 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 60580 00:09:02.665 [2024-11-20 15:57:00.732709] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:02.665 15:57:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 60580 00:09:02.665 [2024-11-20 15:57:00.743438] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:03.231 ************************************ 00:09:03.231 END TEST raid_state_function_test_sb 00:09:03.231 ************************************ 00:09:03.231 15:57:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:03.231 00:09:03.231 real 0m3.639s 00:09:03.231 user 0m5.188s 00:09:03.231 sys 0m0.559s 00:09:03.231 15:57:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:03.231 15:57:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.489 15:57:01 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 2 00:09:03.489 15:57:01 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:09:03.489 15:57:01 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:03.489 15:57:01 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:03.489 ************************************ 00:09:03.489 START TEST raid_superblock_test 00:09:03.489 ************************************ 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 2 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:09:03.489 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=60821 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 60821 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 60821 ']' 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:03.489 15:57:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.489 [2024-11-20 15:57:01.565317] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:09:03.489 [2024-11-20 15:57:01.565423] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid60821 ] 00:09:03.489 [2024-11-20 15:57:01.721141] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:03.747 [2024-11-20 15:57:01.825319] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:03.747 [2024-11-20 15:57:01.964466] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:03.747 [2024-11-20 15:57:01.964557] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.311 malloc1 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.311 [2024-11-20 15:57:02.422638] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:04.311 [2024-11-20 15:57:02.422873] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:04.311 [2024-11-20 15:57:02.422919] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:04.311 [2024-11-20 15:57:02.423133] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:04.311 [2024-11-20 15:57:02.425406] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:04.311 [2024-11-20 15:57:02.425530] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:04.311 pt1 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.311 malloc2 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.311 [2024-11-20 15:57:02.463388] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:04.311 [2024-11-20 15:57:02.463541] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:04.311 [2024-11-20 15:57:02.463571] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:04.311 [2024-11-20 15:57:02.463581] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:04.311 [2024-11-20 15:57:02.465898] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:04.311 [2024-11-20 15:57:02.466005] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:04.311 pt2 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.311 [2024-11-20 15:57:02.475455] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:04.311 [2024-11-20 15:57:02.477391] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:04.311 [2024-11-20 15:57:02.477597] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:09:04.311 [2024-11-20 15:57:02.477617] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:09:04.311 [2024-11-20 15:57:02.477903] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:09:04.311 [2024-11-20 15:57:02.478043] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:09:04.311 [2024-11-20 15:57:02.478054] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:09:04.311 [2024-11-20 15:57:02.478203] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:04.311 "name": "raid_bdev1", 00:09:04.311 "uuid": "61179c00-1182-4167-8792-1f36f0eb23c8", 00:09:04.311 "strip_size_kb": 64, 00:09:04.311 "state": "online", 00:09:04.311 "raid_level": "concat", 00:09:04.311 "superblock": true, 00:09:04.311 "num_base_bdevs": 2, 00:09:04.311 "num_base_bdevs_discovered": 2, 00:09:04.311 "num_base_bdevs_operational": 2, 00:09:04.311 "base_bdevs_list": [ 00:09:04.311 { 00:09:04.311 "name": "pt1", 00:09:04.311 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:04.311 "is_configured": true, 00:09:04.311 "data_offset": 2048, 00:09:04.311 "data_size": 63488 00:09:04.311 }, 00:09:04.311 { 00:09:04.311 "name": "pt2", 00:09:04.311 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:04.311 "is_configured": true, 00:09:04.311 "data_offset": 2048, 00:09:04.311 "data_size": 63488 00:09:04.311 } 00:09:04.311 ] 00:09:04.311 }' 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:04.311 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.568 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:04.568 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:04.568 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:04.568 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:04.568 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:04.568 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:04.568 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:04.568 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:04.568 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.568 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.568 [2024-11-20 15:57:02.795809] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:04.568 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:04.827 "name": "raid_bdev1", 00:09:04.827 "aliases": [ 00:09:04.827 "61179c00-1182-4167-8792-1f36f0eb23c8" 00:09:04.827 ], 00:09:04.827 "product_name": "Raid Volume", 00:09:04.827 "block_size": 512, 00:09:04.827 "num_blocks": 126976, 00:09:04.827 "uuid": "61179c00-1182-4167-8792-1f36f0eb23c8", 00:09:04.827 "assigned_rate_limits": { 00:09:04.827 "rw_ios_per_sec": 0, 00:09:04.827 "rw_mbytes_per_sec": 0, 00:09:04.827 "r_mbytes_per_sec": 0, 00:09:04.827 "w_mbytes_per_sec": 0 00:09:04.827 }, 00:09:04.827 "claimed": false, 00:09:04.827 "zoned": false, 00:09:04.827 "supported_io_types": { 00:09:04.827 "read": true, 00:09:04.827 "write": true, 00:09:04.827 "unmap": true, 00:09:04.827 "flush": true, 00:09:04.827 "reset": true, 00:09:04.827 "nvme_admin": false, 00:09:04.827 "nvme_io": false, 00:09:04.827 "nvme_io_md": false, 00:09:04.827 "write_zeroes": true, 00:09:04.827 "zcopy": false, 00:09:04.827 "get_zone_info": false, 00:09:04.827 "zone_management": false, 00:09:04.827 "zone_append": false, 00:09:04.827 "compare": false, 00:09:04.827 "compare_and_write": false, 00:09:04.827 "abort": false, 00:09:04.827 "seek_hole": false, 00:09:04.827 "seek_data": false, 00:09:04.827 "copy": false, 00:09:04.827 "nvme_iov_md": false 00:09:04.827 }, 00:09:04.827 "memory_domains": [ 00:09:04.827 { 00:09:04.827 "dma_device_id": "system", 00:09:04.827 "dma_device_type": 1 00:09:04.827 }, 00:09:04.827 { 00:09:04.827 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:04.827 "dma_device_type": 2 00:09:04.827 }, 00:09:04.827 { 00:09:04.827 "dma_device_id": "system", 00:09:04.827 "dma_device_type": 1 00:09:04.827 }, 00:09:04.827 { 00:09:04.827 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:04.827 "dma_device_type": 2 00:09:04.827 } 00:09:04.827 ], 00:09:04.827 "driver_specific": { 00:09:04.827 "raid": { 00:09:04.827 "uuid": "61179c00-1182-4167-8792-1f36f0eb23c8", 00:09:04.827 "strip_size_kb": 64, 00:09:04.827 "state": "online", 00:09:04.827 "raid_level": "concat", 00:09:04.827 "superblock": true, 00:09:04.827 "num_base_bdevs": 2, 00:09:04.827 "num_base_bdevs_discovered": 2, 00:09:04.827 "num_base_bdevs_operational": 2, 00:09:04.827 "base_bdevs_list": [ 00:09:04.827 { 00:09:04.827 "name": "pt1", 00:09:04.827 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:04.827 "is_configured": true, 00:09:04.827 "data_offset": 2048, 00:09:04.827 "data_size": 63488 00:09:04.827 }, 00:09:04.827 { 00:09:04.827 "name": "pt2", 00:09:04.827 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:04.827 "is_configured": true, 00:09:04.827 "data_offset": 2048, 00:09:04.827 "data_size": 63488 00:09:04.827 } 00:09:04.827 ] 00:09:04.827 } 00:09:04.827 } 00:09:04.827 }' 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:04.827 pt2' 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.827 [2024-11-20 15:57:02.955847] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=61179c00-1182-4167-8792-1f36f0eb23c8 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 61179c00-1182-4167-8792-1f36f0eb23c8 ']' 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.827 [2024-11-20 15:57:02.991503] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:04.827 [2024-11-20 15:57:02.991641] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:04.827 [2024-11-20 15:57:02.991762] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:04.827 [2024-11-20 15:57:02.991813] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:04.827 [2024-11-20 15:57:02.991824] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.827 15:57:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.827 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.086 [2024-11-20 15:57:03.083590] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:05.086 [2024-11-20 15:57:03.085600] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:05.086 [2024-11-20 15:57:03.085779] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:05.086 [2024-11-20 15:57:03.085837] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:05.086 [2024-11-20 15:57:03.085852] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:05.086 [2024-11-20 15:57:03.085863] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:09:05.086 request: 00:09:05.086 { 00:09:05.086 "name": "raid_bdev1", 00:09:05.086 "raid_level": "concat", 00:09:05.086 "base_bdevs": [ 00:09:05.086 "malloc1", 00:09:05.086 "malloc2" 00:09:05.086 ], 00:09:05.086 "strip_size_kb": 64, 00:09:05.086 "superblock": false, 00:09:05.086 "method": "bdev_raid_create", 00:09:05.086 "req_id": 1 00:09:05.086 } 00:09:05.086 Got JSON-RPC error response 00:09:05.086 response: 00:09:05.086 { 00:09:05.086 "code": -17, 00:09:05.086 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:05.086 } 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.086 [2024-11-20 15:57:03.123584] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:05.086 [2024-11-20 15:57:03.123770] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:05.086 [2024-11-20 15:57:03.123914] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:05.086 [2024-11-20 15:57:03.123984] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:05.086 [2024-11-20 15:57:03.126233] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:05.086 [2024-11-20 15:57:03.126354] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:05.086 pt1 00:09:05.086 [2024-11-20 15:57:03.126483] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:05.086 [2024-11-20 15:57:03.126538] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 2 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.086 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:05.086 "name": "raid_bdev1", 00:09:05.086 "uuid": "61179c00-1182-4167-8792-1f36f0eb23c8", 00:09:05.086 "strip_size_kb": 64, 00:09:05.086 "state": "configuring", 00:09:05.086 "raid_level": "concat", 00:09:05.086 "superblock": true, 00:09:05.086 "num_base_bdevs": 2, 00:09:05.086 "num_base_bdevs_discovered": 1, 00:09:05.086 "num_base_bdevs_operational": 2, 00:09:05.086 "base_bdevs_list": [ 00:09:05.086 { 00:09:05.086 "name": "pt1", 00:09:05.086 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:05.086 "is_configured": true, 00:09:05.086 "data_offset": 2048, 00:09:05.086 "data_size": 63488 00:09:05.086 }, 00:09:05.086 { 00:09:05.086 "name": null, 00:09:05.087 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:05.087 "is_configured": false, 00:09:05.087 "data_offset": 2048, 00:09:05.087 "data_size": 63488 00:09:05.087 } 00:09:05.087 ] 00:09:05.087 }' 00:09:05.087 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:05.087 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.417 [2024-11-20 15:57:03.439680] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:05.417 [2024-11-20 15:57:03.439869] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:05.417 [2024-11-20 15:57:03.439909] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:09:05.417 [2024-11-20 15:57:03.440019] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:05.417 [2024-11-20 15:57:03.440436] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:05.417 [2024-11-20 15:57:03.440463] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:05.417 [2024-11-20 15:57:03.440536] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:05.417 [2024-11-20 15:57:03.440559] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:05.417 [2024-11-20 15:57:03.440661] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:05.417 [2024-11-20 15:57:03.440687] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:09:05.417 [2024-11-20 15:57:03.440981] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:09:05.417 [2024-11-20 15:57:03.441162] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:05.417 [2024-11-20 15:57:03.441174] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:09:05.417 [2024-11-20 15:57:03.441301] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:05.417 pt2 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.417 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:05.417 "name": "raid_bdev1", 00:09:05.417 "uuid": "61179c00-1182-4167-8792-1f36f0eb23c8", 00:09:05.417 "strip_size_kb": 64, 00:09:05.418 "state": "online", 00:09:05.418 "raid_level": "concat", 00:09:05.418 "superblock": true, 00:09:05.418 "num_base_bdevs": 2, 00:09:05.418 "num_base_bdevs_discovered": 2, 00:09:05.418 "num_base_bdevs_operational": 2, 00:09:05.418 "base_bdevs_list": [ 00:09:05.418 { 00:09:05.418 "name": "pt1", 00:09:05.418 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:05.418 "is_configured": true, 00:09:05.418 "data_offset": 2048, 00:09:05.418 "data_size": 63488 00:09:05.418 }, 00:09:05.418 { 00:09:05.418 "name": "pt2", 00:09:05.418 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:05.418 "is_configured": true, 00:09:05.418 "data_offset": 2048, 00:09:05.418 "data_size": 63488 00:09:05.418 } 00:09:05.418 ] 00:09:05.418 }' 00:09:05.418 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:05.418 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.685 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:05.685 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:05.685 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:05.685 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:05.685 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:05.685 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:05.685 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:05.685 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:05.685 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.685 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.685 [2024-11-20 15:57:03.743983] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:05.685 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.685 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:05.685 "name": "raid_bdev1", 00:09:05.685 "aliases": [ 00:09:05.685 "61179c00-1182-4167-8792-1f36f0eb23c8" 00:09:05.685 ], 00:09:05.685 "product_name": "Raid Volume", 00:09:05.685 "block_size": 512, 00:09:05.685 "num_blocks": 126976, 00:09:05.685 "uuid": "61179c00-1182-4167-8792-1f36f0eb23c8", 00:09:05.685 "assigned_rate_limits": { 00:09:05.685 "rw_ios_per_sec": 0, 00:09:05.685 "rw_mbytes_per_sec": 0, 00:09:05.685 "r_mbytes_per_sec": 0, 00:09:05.685 "w_mbytes_per_sec": 0 00:09:05.685 }, 00:09:05.685 "claimed": false, 00:09:05.686 "zoned": false, 00:09:05.686 "supported_io_types": { 00:09:05.686 "read": true, 00:09:05.686 "write": true, 00:09:05.686 "unmap": true, 00:09:05.686 "flush": true, 00:09:05.686 "reset": true, 00:09:05.686 "nvme_admin": false, 00:09:05.686 "nvme_io": false, 00:09:05.686 "nvme_io_md": false, 00:09:05.686 "write_zeroes": true, 00:09:05.686 "zcopy": false, 00:09:05.686 "get_zone_info": false, 00:09:05.686 "zone_management": false, 00:09:05.686 "zone_append": false, 00:09:05.686 "compare": false, 00:09:05.686 "compare_and_write": false, 00:09:05.686 "abort": false, 00:09:05.686 "seek_hole": false, 00:09:05.686 "seek_data": false, 00:09:05.686 "copy": false, 00:09:05.686 "nvme_iov_md": false 00:09:05.686 }, 00:09:05.686 "memory_domains": [ 00:09:05.686 { 00:09:05.686 "dma_device_id": "system", 00:09:05.686 "dma_device_type": 1 00:09:05.686 }, 00:09:05.686 { 00:09:05.686 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:05.686 "dma_device_type": 2 00:09:05.686 }, 00:09:05.686 { 00:09:05.686 "dma_device_id": "system", 00:09:05.686 "dma_device_type": 1 00:09:05.686 }, 00:09:05.686 { 00:09:05.686 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:05.686 "dma_device_type": 2 00:09:05.686 } 00:09:05.686 ], 00:09:05.686 "driver_specific": { 00:09:05.686 "raid": { 00:09:05.686 "uuid": "61179c00-1182-4167-8792-1f36f0eb23c8", 00:09:05.686 "strip_size_kb": 64, 00:09:05.686 "state": "online", 00:09:05.686 "raid_level": "concat", 00:09:05.686 "superblock": true, 00:09:05.686 "num_base_bdevs": 2, 00:09:05.686 "num_base_bdevs_discovered": 2, 00:09:05.686 "num_base_bdevs_operational": 2, 00:09:05.686 "base_bdevs_list": [ 00:09:05.686 { 00:09:05.686 "name": "pt1", 00:09:05.686 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:05.686 "is_configured": true, 00:09:05.686 "data_offset": 2048, 00:09:05.686 "data_size": 63488 00:09:05.686 }, 00:09:05.686 { 00:09:05.686 "name": "pt2", 00:09:05.686 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:05.686 "is_configured": true, 00:09:05.686 "data_offset": 2048, 00:09:05.686 "data_size": 63488 00:09:05.686 } 00:09:05.686 ] 00:09:05.686 } 00:09:05.686 } 00:09:05.686 }' 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:05.686 pt2' 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.686 [2024-11-20 15:57:03.908032] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 61179c00-1182-4167-8792-1f36f0eb23c8 '!=' 61179c00-1182-4167-8792-1f36f0eb23c8 ']' 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 60821 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 60821 ']' 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 60821 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:09:05.686 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:05.944 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 60821 00:09:05.944 killing process with pid 60821 00:09:05.944 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:05.944 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:05.944 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 60821' 00:09:05.944 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 60821 00:09:05.944 [2024-11-20 15:57:03.951903] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:05.944 15:57:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 60821 00:09:05.944 [2024-11-20 15:57:03.951984] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:05.944 [2024-11-20 15:57:03.952031] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:05.944 [2024-11-20 15:57:03.952045] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:09:05.944 [2024-11-20 15:57:04.084724] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:06.875 15:57:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:06.875 00:09:06.875 real 0m3.280s 00:09:06.875 user 0m4.609s 00:09:06.875 sys 0m0.465s 00:09:06.875 ************************************ 00:09:06.875 END TEST raid_superblock_test 00:09:06.875 ************************************ 00:09:06.875 15:57:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:06.876 15:57:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.876 15:57:04 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 2 read 00:09:06.876 15:57:04 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:06.876 15:57:04 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:06.876 15:57:04 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:06.876 ************************************ 00:09:06.876 START TEST raid_read_error_test 00:09:06.876 ************************************ 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 2 read 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.47WVMVVTiy 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=61016 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 61016 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 61016 ']' 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:06.876 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:06.876 15:57:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.876 [2024-11-20 15:57:04.911085] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:09:06.876 [2024-11-20 15:57:04.911214] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid61016 ] 00:09:06.876 [2024-11-20 15:57:05.072814] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:07.134 [2024-11-20 15:57:05.159228] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:07.134 [2024-11-20 15:57:05.272492] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:07.134 [2024-11-20 15:57:05.272542] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.702 BaseBdev1_malloc 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.702 true 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.702 [2024-11-20 15:57:05.867577] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:07.702 [2024-11-20 15:57:05.867745] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:07.702 [2024-11-20 15:57:05.867783] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:07.702 [2024-11-20 15:57:05.867988] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:07.702 [2024-11-20 15:57:05.869833] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:07.702 [2024-11-20 15:57:05.869861] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:07.702 BaseBdev1 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.702 BaseBdev2_malloc 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.702 true 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.702 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.703 [2024-11-20 15:57:05.912240] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:07.703 [2024-11-20 15:57:05.912405] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:07.703 [2024-11-20 15:57:05.912441] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:07.703 [2024-11-20 15:57:05.912617] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:07.703 [2024-11-20 15:57:05.914478] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:07.703 [2024-11-20 15:57:05.914512] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:07.703 BaseBdev2 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.703 [2024-11-20 15:57:05.920278] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:07.703 [2024-11-20 15:57:05.921978] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:07.703 [2024-11-20 15:57:05.922210] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:07.703 [2024-11-20 15:57:05.922277] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:09:07.703 [2024-11-20 15:57:05.922550] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:09:07.703 [2024-11-20 15:57:05.922753] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:07.703 [2024-11-20 15:57:05.922824] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:09:07.703 [2024-11-20 15:57:05.923005] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.703 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.962 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:07.962 "name": "raid_bdev1", 00:09:07.962 "uuid": "40e9165a-b0ae-4f01-a820-fcecbf3c46f5", 00:09:07.962 "strip_size_kb": 64, 00:09:07.962 "state": "online", 00:09:07.962 "raid_level": "concat", 00:09:07.962 "superblock": true, 00:09:07.962 "num_base_bdevs": 2, 00:09:07.962 "num_base_bdevs_discovered": 2, 00:09:07.962 "num_base_bdevs_operational": 2, 00:09:07.962 "base_bdevs_list": [ 00:09:07.962 { 00:09:07.962 "name": "BaseBdev1", 00:09:07.962 "uuid": "b89f05f6-02a8-56a5-a4c3-37aebe001a9a", 00:09:07.962 "is_configured": true, 00:09:07.962 "data_offset": 2048, 00:09:07.962 "data_size": 63488 00:09:07.962 }, 00:09:07.962 { 00:09:07.962 "name": "BaseBdev2", 00:09:07.962 "uuid": "09884f3d-03dd-5e95-8f84-80fb394a67b2", 00:09:07.962 "is_configured": true, 00:09:07.962 "data_offset": 2048, 00:09:07.962 "data_size": 63488 00:09:07.962 } 00:09:07.962 ] 00:09:07.962 }' 00:09:07.962 15:57:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:07.962 15:57:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.219 15:57:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:08.219 15:57:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:08.219 [2024-11-20 15:57:06.329130] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:09.152 "name": "raid_bdev1", 00:09:09.152 "uuid": "40e9165a-b0ae-4f01-a820-fcecbf3c46f5", 00:09:09.152 "strip_size_kb": 64, 00:09:09.152 "state": "online", 00:09:09.152 "raid_level": "concat", 00:09:09.152 "superblock": true, 00:09:09.152 "num_base_bdevs": 2, 00:09:09.152 "num_base_bdevs_discovered": 2, 00:09:09.152 "num_base_bdevs_operational": 2, 00:09:09.152 "base_bdevs_list": [ 00:09:09.152 { 00:09:09.152 "name": "BaseBdev1", 00:09:09.152 "uuid": "b89f05f6-02a8-56a5-a4c3-37aebe001a9a", 00:09:09.152 "is_configured": true, 00:09:09.152 "data_offset": 2048, 00:09:09.152 "data_size": 63488 00:09:09.152 }, 00:09:09.152 { 00:09:09.152 "name": "BaseBdev2", 00:09:09.152 "uuid": "09884f3d-03dd-5e95-8f84-80fb394a67b2", 00:09:09.152 "is_configured": true, 00:09:09.152 "data_offset": 2048, 00:09:09.152 "data_size": 63488 00:09:09.152 } 00:09:09.152 ] 00:09:09.152 }' 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:09.152 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.470 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:09.470 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.470 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.470 [2024-11-20 15:57:07.573994] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:09.470 [2024-11-20 15:57:07.574025] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:09.470 [2024-11-20 15:57:07.576547] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:09.470 [2024-11-20 15:57:07.576585] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:09.470 [2024-11-20 15:57:07.576611] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:09.470 [2024-11-20 15:57:07.576622] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:09:09.470 { 00:09:09.470 "results": [ 00:09:09.470 { 00:09:09.470 "job": "raid_bdev1", 00:09:09.470 "core_mask": "0x1", 00:09:09.470 "workload": "randrw", 00:09:09.470 "percentage": 50, 00:09:09.470 "status": "finished", 00:09:09.470 "queue_depth": 1, 00:09:09.470 "io_size": 131072, 00:09:09.470 "runtime": 1.243391, 00:09:09.470 "iops": 17520.635101910822, 00:09:09.470 "mibps": 2190.079387738853, 00:09:09.470 "io_failed": 1, 00:09:09.470 "io_timeout": 0, 00:09:09.470 "avg_latency_us": 78.10325275935853, 00:09:09.470 "min_latency_us": 25.796923076923076, 00:09:09.470 "max_latency_us": 1348.5292307692307 00:09:09.470 } 00:09:09.470 ], 00:09:09.470 "core_count": 1 00:09:09.470 } 00:09:09.470 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.470 15:57:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 61016 00:09:09.470 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 61016 ']' 00:09:09.470 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 61016 00:09:09.470 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:09:09.470 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:09.470 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 61016 00:09:09.470 killing process with pid 61016 00:09:09.470 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:09.470 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:09.470 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 61016' 00:09:09.470 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 61016 00:09:09.470 [2024-11-20 15:57:07.609696] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:09.470 15:57:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 61016 00:09:09.470 [2024-11-20 15:57:07.677809] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:10.414 15:57:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:10.414 15:57:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.47WVMVVTiy 00:09:10.414 15:57:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:10.414 ************************************ 00:09:10.414 END TEST raid_read_error_test 00:09:10.414 ************************************ 00:09:10.414 15:57:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.80 00:09:10.414 15:57:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:09:10.414 15:57:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:10.414 15:57:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:10.414 15:57:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.80 != \0\.\0\0 ]] 00:09:10.414 00:09:10.414 real 0m3.479s 00:09:10.414 user 0m4.252s 00:09:10.414 sys 0m0.400s 00:09:10.414 15:57:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:10.414 15:57:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.414 15:57:08 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 2 write 00:09:10.414 15:57:08 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:10.414 15:57:08 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:10.414 15:57:08 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:10.414 ************************************ 00:09:10.414 START TEST raid_write_error_test 00:09:10.414 ************************************ 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 2 write 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:10.414 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.LZVA7Jol1A 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=61153 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 61153 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 61153 ']' 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:10.414 15:57:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.414 [2024-11-20 15:57:08.439516] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:09:10.414 [2024-11-20 15:57:08.439708] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid61153 ] 00:09:10.414 [2024-11-20 15:57:08.609833] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:10.672 [2024-11-20 15:57:08.695102] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:10.672 [2024-11-20 15:57:08.806818] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:10.672 [2024-11-20 15:57:08.806854] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:11.239 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.240 BaseBdev1_malloc 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.240 true 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.240 [2024-11-20 15:57:09.226612] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:11.240 [2024-11-20 15:57:09.226782] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:11.240 [2024-11-20 15:57:09.226819] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:11.240 [2024-11-20 15:57:09.226830] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:11.240 [2024-11-20 15:57:09.228622] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:11.240 [2024-11-20 15:57:09.228655] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:11.240 BaseBdev1 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.240 BaseBdev2_malloc 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.240 true 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.240 [2024-11-20 15:57:09.266112] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:11.240 [2024-11-20 15:57:09.266243] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:11.240 [2024-11-20 15:57:09.266278] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:11.240 [2024-11-20 15:57:09.266354] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:11.240 [2024-11-20 15:57:09.268144] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:11.240 [2024-11-20 15:57:09.268242] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:11.240 BaseBdev2 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.240 [2024-11-20 15:57:09.274169] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:11.240 [2024-11-20 15:57:09.275789] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:11.240 [2024-11-20 15:57:09.276014] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:11.240 [2024-11-20 15:57:09.276080] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:09:11.240 [2024-11-20 15:57:09.276306] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:09:11.240 [2024-11-20 15:57:09.276483] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:11.240 [2024-11-20 15:57:09.276542] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:09:11.240 [2024-11-20 15:57:09.276740] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:11.240 "name": "raid_bdev1", 00:09:11.240 "uuid": "f6e95e9d-0fc3-4024-84b4-315c2c361776", 00:09:11.240 "strip_size_kb": 64, 00:09:11.240 "state": "online", 00:09:11.240 "raid_level": "concat", 00:09:11.240 "superblock": true, 00:09:11.240 "num_base_bdevs": 2, 00:09:11.240 "num_base_bdevs_discovered": 2, 00:09:11.240 "num_base_bdevs_operational": 2, 00:09:11.240 "base_bdevs_list": [ 00:09:11.240 { 00:09:11.240 "name": "BaseBdev1", 00:09:11.240 "uuid": "027a3ad7-a973-552f-8bfb-50f174fb2afc", 00:09:11.240 "is_configured": true, 00:09:11.240 "data_offset": 2048, 00:09:11.240 "data_size": 63488 00:09:11.240 }, 00:09:11.240 { 00:09:11.240 "name": "BaseBdev2", 00:09:11.240 "uuid": "6e72e322-e588-52e5-bc27-ca2f39089c3b", 00:09:11.240 "is_configured": true, 00:09:11.240 "data_offset": 2048, 00:09:11.240 "data_size": 63488 00:09:11.240 } 00:09:11.240 ] 00:09:11.240 }' 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:11.240 15:57:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.498 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:11.498 15:57:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:11.498 [2024-11-20 15:57:09.667007] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:12.431 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.432 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.432 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:12.432 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.432 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.432 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:12.432 "name": "raid_bdev1", 00:09:12.432 "uuid": "f6e95e9d-0fc3-4024-84b4-315c2c361776", 00:09:12.432 "strip_size_kb": 64, 00:09:12.432 "state": "online", 00:09:12.432 "raid_level": "concat", 00:09:12.432 "superblock": true, 00:09:12.432 "num_base_bdevs": 2, 00:09:12.432 "num_base_bdevs_discovered": 2, 00:09:12.432 "num_base_bdevs_operational": 2, 00:09:12.432 "base_bdevs_list": [ 00:09:12.432 { 00:09:12.432 "name": "BaseBdev1", 00:09:12.432 "uuid": "027a3ad7-a973-552f-8bfb-50f174fb2afc", 00:09:12.432 "is_configured": true, 00:09:12.432 "data_offset": 2048, 00:09:12.432 "data_size": 63488 00:09:12.432 }, 00:09:12.432 { 00:09:12.432 "name": "BaseBdev2", 00:09:12.432 "uuid": "6e72e322-e588-52e5-bc27-ca2f39089c3b", 00:09:12.432 "is_configured": true, 00:09:12.432 "data_offset": 2048, 00:09:12.432 "data_size": 63488 00:09:12.432 } 00:09:12.432 ] 00:09:12.432 }' 00:09:12.432 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:12.432 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.690 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:12.690 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.690 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.690 [2024-11-20 15:57:10.936050] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:12.690 [2024-11-20 15:57:10.936172] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:12.690 [2024-11-20 15:57:10.938654] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:12.948 [2024-11-20 15:57:10.938781] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:12.948 [2024-11-20 15:57:10.938825] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:12.948 [2024-11-20 15:57:10.938930] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:09:12.948 { 00:09:12.948 "results": [ 00:09:12.948 { 00:09:12.948 "job": "raid_bdev1", 00:09:12.948 "core_mask": "0x1", 00:09:12.948 "workload": "randrw", 00:09:12.948 "percentage": 50, 00:09:12.948 "status": "finished", 00:09:12.948 "queue_depth": 1, 00:09:12.948 "io_size": 131072, 00:09:12.948 "runtime": 1.267629, 00:09:12.948 "iops": 18295.573862699577, 00:09:12.948 "mibps": 2286.946732837447, 00:09:12.948 "io_failed": 1, 00:09:12.948 "io_timeout": 0, 00:09:12.948 "avg_latency_us": 74.6697784809077, 00:09:12.948 "min_latency_us": 25.993846153846153, 00:09:12.948 "max_latency_us": 1317.0215384615385 00:09:12.948 } 00:09:12.948 ], 00:09:12.948 "core_count": 1 00:09:12.948 } 00:09:12.948 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.948 15:57:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 61153 00:09:12.948 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 61153 ']' 00:09:12.948 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 61153 00:09:12.948 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:09:12.948 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:12.948 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 61153 00:09:12.948 killing process with pid 61153 00:09:12.948 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:12.948 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:12.948 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 61153' 00:09:12.948 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 61153 00:09:12.948 [2024-11-20 15:57:10.969210] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:12.948 15:57:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 61153 00:09:12.948 [2024-11-20 15:57:11.036105] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:13.514 15:57:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.LZVA7Jol1A 00:09:13.514 15:57:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:13.514 15:57:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:13.514 15:57:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.79 00:09:13.514 15:57:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:09:13.514 15:57:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:13.514 15:57:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:13.514 ************************************ 00:09:13.514 END TEST raid_write_error_test 00:09:13.514 ************************************ 00:09:13.514 15:57:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.79 != \0\.\0\0 ]] 00:09:13.514 00:09:13.514 real 0m3.302s 00:09:13.514 user 0m3.939s 00:09:13.514 sys 0m0.375s 00:09:13.514 15:57:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:13.514 15:57:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.514 15:57:11 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:13.514 15:57:11 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 2 false 00:09:13.514 15:57:11 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:13.514 15:57:11 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:13.514 15:57:11 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:13.514 ************************************ 00:09:13.514 START TEST raid_state_function_test 00:09:13.514 ************************************ 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 false 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:13.514 Process raid pid: 61280 00:09:13.514 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=61280 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 61280' 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 61280 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 61280 ']' 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.514 15:57:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:13.514 [2024-11-20 15:57:11.757905] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:09:13.514 [2024-11-20 15:57:11.758021] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:13.771 [2024-11-20 15:57:11.911323] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:13.771 [2024-11-20 15:57:11.994800] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:14.029 [2024-11-20 15:57:12.103434] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:14.029 [2024-11-20 15:57:12.103468] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.287 [2024-11-20 15:57:12.522056] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:14.287 [2024-11-20 15:57:12.522206] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:14.287 [2024-11-20 15:57:12.522264] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:14.287 [2024-11-20 15:57:12.522288] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.287 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:14.544 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.544 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:14.544 "name": "Existed_Raid", 00:09:14.544 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:14.544 "strip_size_kb": 0, 00:09:14.544 "state": "configuring", 00:09:14.544 "raid_level": "raid1", 00:09:14.544 "superblock": false, 00:09:14.544 "num_base_bdevs": 2, 00:09:14.544 "num_base_bdevs_discovered": 0, 00:09:14.544 "num_base_bdevs_operational": 2, 00:09:14.544 "base_bdevs_list": [ 00:09:14.544 { 00:09:14.544 "name": "BaseBdev1", 00:09:14.544 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:14.544 "is_configured": false, 00:09:14.544 "data_offset": 0, 00:09:14.544 "data_size": 0 00:09:14.544 }, 00:09:14.544 { 00:09:14.544 "name": "BaseBdev2", 00:09:14.544 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:14.544 "is_configured": false, 00:09:14.544 "data_offset": 0, 00:09:14.544 "data_size": 0 00:09:14.544 } 00:09:14.544 ] 00:09:14.544 }' 00:09:14.544 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:14.544 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.802 [2024-11-20 15:57:12.838092] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:14.802 [2024-11-20 15:57:12.838219] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.802 [2024-11-20 15:57:12.846078] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:14.802 [2024-11-20 15:57:12.846183] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:14.802 [2024-11-20 15:57:12.846234] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:14.802 [2024-11-20 15:57:12.846256] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.802 [2024-11-20 15:57:12.874460] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:14.802 BaseBdev1 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.802 [ 00:09:14.802 { 00:09:14.802 "name": "BaseBdev1", 00:09:14.802 "aliases": [ 00:09:14.802 "cb2184b7-f6c4-4ca9-8ac6-4852bbd9b52b" 00:09:14.802 ], 00:09:14.802 "product_name": "Malloc disk", 00:09:14.802 "block_size": 512, 00:09:14.802 "num_blocks": 65536, 00:09:14.802 "uuid": "cb2184b7-f6c4-4ca9-8ac6-4852bbd9b52b", 00:09:14.802 "assigned_rate_limits": { 00:09:14.802 "rw_ios_per_sec": 0, 00:09:14.802 "rw_mbytes_per_sec": 0, 00:09:14.802 "r_mbytes_per_sec": 0, 00:09:14.802 "w_mbytes_per_sec": 0 00:09:14.802 }, 00:09:14.802 "claimed": true, 00:09:14.802 "claim_type": "exclusive_write", 00:09:14.802 "zoned": false, 00:09:14.802 "supported_io_types": { 00:09:14.802 "read": true, 00:09:14.802 "write": true, 00:09:14.802 "unmap": true, 00:09:14.802 "flush": true, 00:09:14.802 "reset": true, 00:09:14.802 "nvme_admin": false, 00:09:14.802 "nvme_io": false, 00:09:14.802 "nvme_io_md": false, 00:09:14.802 "write_zeroes": true, 00:09:14.802 "zcopy": true, 00:09:14.802 "get_zone_info": false, 00:09:14.802 "zone_management": false, 00:09:14.802 "zone_append": false, 00:09:14.802 "compare": false, 00:09:14.802 "compare_and_write": false, 00:09:14.802 "abort": true, 00:09:14.802 "seek_hole": false, 00:09:14.802 "seek_data": false, 00:09:14.802 "copy": true, 00:09:14.802 "nvme_iov_md": false 00:09:14.802 }, 00:09:14.802 "memory_domains": [ 00:09:14.802 { 00:09:14.802 "dma_device_id": "system", 00:09:14.802 "dma_device_type": 1 00:09:14.802 }, 00:09:14.802 { 00:09:14.802 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:14.802 "dma_device_type": 2 00:09:14.802 } 00:09:14.802 ], 00:09:14.802 "driver_specific": {} 00:09:14.802 } 00:09:14.802 ] 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:14.802 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:09:14.803 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:14.803 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:14.803 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:14.803 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:14.803 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:14.803 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:14.803 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:14.803 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:14.803 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:14.803 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.803 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:14.803 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.803 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.803 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.803 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:14.803 "name": "Existed_Raid", 00:09:14.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:14.803 "strip_size_kb": 0, 00:09:14.803 "state": "configuring", 00:09:14.803 "raid_level": "raid1", 00:09:14.803 "superblock": false, 00:09:14.803 "num_base_bdevs": 2, 00:09:14.803 "num_base_bdevs_discovered": 1, 00:09:14.803 "num_base_bdevs_operational": 2, 00:09:14.803 "base_bdevs_list": [ 00:09:14.803 { 00:09:14.803 "name": "BaseBdev1", 00:09:14.803 "uuid": "cb2184b7-f6c4-4ca9-8ac6-4852bbd9b52b", 00:09:14.803 "is_configured": true, 00:09:14.803 "data_offset": 0, 00:09:14.803 "data_size": 65536 00:09:14.803 }, 00:09:14.803 { 00:09:14.803 "name": "BaseBdev2", 00:09:14.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:14.803 "is_configured": false, 00:09:14.803 "data_offset": 0, 00:09:14.803 "data_size": 0 00:09:14.803 } 00:09:14.803 ] 00:09:14.803 }' 00:09:14.803 15:57:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:14.803 15:57:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.060 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:15.060 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.060 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.060 [2024-11-20 15:57:13.186558] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:15.060 [2024-11-20 15:57:13.186686] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:09:15.060 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.060 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:09:15.060 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.060 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.060 [2024-11-20 15:57:13.194590] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:15.060 [2024-11-20 15:57:13.196208] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:15.060 [2024-11-20 15:57:13.196311] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:15.060 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.060 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:15.061 "name": "Existed_Raid", 00:09:15.061 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:15.061 "strip_size_kb": 0, 00:09:15.061 "state": "configuring", 00:09:15.061 "raid_level": "raid1", 00:09:15.061 "superblock": false, 00:09:15.061 "num_base_bdevs": 2, 00:09:15.061 "num_base_bdevs_discovered": 1, 00:09:15.061 "num_base_bdevs_operational": 2, 00:09:15.061 "base_bdevs_list": [ 00:09:15.061 { 00:09:15.061 "name": "BaseBdev1", 00:09:15.061 "uuid": "cb2184b7-f6c4-4ca9-8ac6-4852bbd9b52b", 00:09:15.061 "is_configured": true, 00:09:15.061 "data_offset": 0, 00:09:15.061 "data_size": 65536 00:09:15.061 }, 00:09:15.061 { 00:09:15.061 "name": "BaseBdev2", 00:09:15.061 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:15.061 "is_configured": false, 00:09:15.061 "data_offset": 0, 00:09:15.061 "data_size": 0 00:09:15.061 } 00:09:15.061 ] 00:09:15.061 }' 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:15.061 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.319 BaseBdev2 00:09:15.319 [2024-11-20 15:57:13.520930] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:15.319 [2024-11-20 15:57:13.520971] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:15.319 [2024-11-20 15:57:13.520978] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:09:15.319 [2024-11-20 15:57:13.521183] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:09:15.319 [2024-11-20 15:57:13.521307] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:15.319 [2024-11-20 15:57:13.521317] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:09:15.319 [2024-11-20 15:57:13.521509] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.319 [ 00:09:15.319 { 00:09:15.319 "name": "BaseBdev2", 00:09:15.319 "aliases": [ 00:09:15.319 "b68d43bc-9aba-4d47-b12c-61661c0dfa03" 00:09:15.319 ], 00:09:15.319 "product_name": "Malloc disk", 00:09:15.319 "block_size": 512, 00:09:15.319 "num_blocks": 65536, 00:09:15.319 "uuid": "b68d43bc-9aba-4d47-b12c-61661c0dfa03", 00:09:15.319 "assigned_rate_limits": { 00:09:15.319 "rw_ios_per_sec": 0, 00:09:15.319 "rw_mbytes_per_sec": 0, 00:09:15.319 "r_mbytes_per_sec": 0, 00:09:15.319 "w_mbytes_per_sec": 0 00:09:15.319 }, 00:09:15.319 "claimed": true, 00:09:15.319 "claim_type": "exclusive_write", 00:09:15.319 "zoned": false, 00:09:15.319 "supported_io_types": { 00:09:15.319 "read": true, 00:09:15.319 "write": true, 00:09:15.319 "unmap": true, 00:09:15.319 "flush": true, 00:09:15.319 "reset": true, 00:09:15.319 "nvme_admin": false, 00:09:15.319 "nvme_io": false, 00:09:15.319 "nvme_io_md": false, 00:09:15.319 "write_zeroes": true, 00:09:15.319 "zcopy": true, 00:09:15.319 "get_zone_info": false, 00:09:15.319 "zone_management": false, 00:09:15.319 "zone_append": false, 00:09:15.319 "compare": false, 00:09:15.319 "compare_and_write": false, 00:09:15.319 "abort": true, 00:09:15.319 "seek_hole": false, 00:09:15.319 "seek_data": false, 00:09:15.319 "copy": true, 00:09:15.319 "nvme_iov_md": false 00:09:15.319 }, 00:09:15.319 "memory_domains": [ 00:09:15.319 { 00:09:15.319 "dma_device_id": "system", 00:09:15.319 "dma_device_type": 1 00:09:15.319 }, 00:09:15.319 { 00:09:15.319 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:15.319 "dma_device_type": 2 00:09:15.319 } 00:09:15.319 ], 00:09:15.319 "driver_specific": {} 00:09:15.319 } 00:09:15.319 ] 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:09:15.319 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:15.320 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:15.320 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:15.320 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:15.320 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:15.320 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:15.320 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:15.320 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:15.320 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:15.320 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:15.320 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:15.320 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.320 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.320 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.577 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:15.577 "name": "Existed_Raid", 00:09:15.577 "uuid": "3455e8cc-fde8-44fe-a9af-1a249458b960", 00:09:15.577 "strip_size_kb": 0, 00:09:15.577 "state": "online", 00:09:15.577 "raid_level": "raid1", 00:09:15.577 "superblock": false, 00:09:15.577 "num_base_bdevs": 2, 00:09:15.577 "num_base_bdevs_discovered": 2, 00:09:15.577 "num_base_bdevs_operational": 2, 00:09:15.577 "base_bdevs_list": [ 00:09:15.577 { 00:09:15.577 "name": "BaseBdev1", 00:09:15.577 "uuid": "cb2184b7-f6c4-4ca9-8ac6-4852bbd9b52b", 00:09:15.577 "is_configured": true, 00:09:15.577 "data_offset": 0, 00:09:15.577 "data_size": 65536 00:09:15.577 }, 00:09:15.577 { 00:09:15.577 "name": "BaseBdev2", 00:09:15.577 "uuid": "b68d43bc-9aba-4d47-b12c-61661c0dfa03", 00:09:15.577 "is_configured": true, 00:09:15.577 "data_offset": 0, 00:09:15.577 "data_size": 65536 00:09:15.577 } 00:09:15.577 ] 00:09:15.577 }' 00:09:15.577 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:15.577 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.835 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:15.835 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:15.835 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:15.835 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:15.835 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:15.835 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:15.835 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:15.835 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:15.835 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.835 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.835 [2024-11-20 15:57:13.865269] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:15.835 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:15.836 "name": "Existed_Raid", 00:09:15.836 "aliases": [ 00:09:15.836 "3455e8cc-fde8-44fe-a9af-1a249458b960" 00:09:15.836 ], 00:09:15.836 "product_name": "Raid Volume", 00:09:15.836 "block_size": 512, 00:09:15.836 "num_blocks": 65536, 00:09:15.836 "uuid": "3455e8cc-fde8-44fe-a9af-1a249458b960", 00:09:15.836 "assigned_rate_limits": { 00:09:15.836 "rw_ios_per_sec": 0, 00:09:15.836 "rw_mbytes_per_sec": 0, 00:09:15.836 "r_mbytes_per_sec": 0, 00:09:15.836 "w_mbytes_per_sec": 0 00:09:15.836 }, 00:09:15.836 "claimed": false, 00:09:15.836 "zoned": false, 00:09:15.836 "supported_io_types": { 00:09:15.836 "read": true, 00:09:15.836 "write": true, 00:09:15.836 "unmap": false, 00:09:15.836 "flush": false, 00:09:15.836 "reset": true, 00:09:15.836 "nvme_admin": false, 00:09:15.836 "nvme_io": false, 00:09:15.836 "nvme_io_md": false, 00:09:15.836 "write_zeroes": true, 00:09:15.836 "zcopy": false, 00:09:15.836 "get_zone_info": false, 00:09:15.836 "zone_management": false, 00:09:15.836 "zone_append": false, 00:09:15.836 "compare": false, 00:09:15.836 "compare_and_write": false, 00:09:15.836 "abort": false, 00:09:15.836 "seek_hole": false, 00:09:15.836 "seek_data": false, 00:09:15.836 "copy": false, 00:09:15.836 "nvme_iov_md": false 00:09:15.836 }, 00:09:15.836 "memory_domains": [ 00:09:15.836 { 00:09:15.836 "dma_device_id": "system", 00:09:15.836 "dma_device_type": 1 00:09:15.836 }, 00:09:15.836 { 00:09:15.836 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:15.836 "dma_device_type": 2 00:09:15.836 }, 00:09:15.836 { 00:09:15.836 "dma_device_id": "system", 00:09:15.836 "dma_device_type": 1 00:09:15.836 }, 00:09:15.836 { 00:09:15.836 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:15.836 "dma_device_type": 2 00:09:15.836 } 00:09:15.836 ], 00:09:15.836 "driver_specific": { 00:09:15.836 "raid": { 00:09:15.836 "uuid": "3455e8cc-fde8-44fe-a9af-1a249458b960", 00:09:15.836 "strip_size_kb": 0, 00:09:15.836 "state": "online", 00:09:15.836 "raid_level": "raid1", 00:09:15.836 "superblock": false, 00:09:15.836 "num_base_bdevs": 2, 00:09:15.836 "num_base_bdevs_discovered": 2, 00:09:15.836 "num_base_bdevs_operational": 2, 00:09:15.836 "base_bdevs_list": [ 00:09:15.836 { 00:09:15.836 "name": "BaseBdev1", 00:09:15.836 "uuid": "cb2184b7-f6c4-4ca9-8ac6-4852bbd9b52b", 00:09:15.836 "is_configured": true, 00:09:15.836 "data_offset": 0, 00:09:15.836 "data_size": 65536 00:09:15.836 }, 00:09:15.836 { 00:09:15.836 "name": "BaseBdev2", 00:09:15.836 "uuid": "b68d43bc-9aba-4d47-b12c-61661c0dfa03", 00:09:15.836 "is_configured": true, 00:09:15.836 "data_offset": 0, 00:09:15.836 "data_size": 65536 00:09:15.836 } 00:09:15.836 ] 00:09:15.836 } 00:09:15.836 } 00:09:15.836 }' 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:15.836 BaseBdev2' 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.836 15:57:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.836 [2024-11-20 15:57:14.021099] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.836 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:16.094 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.094 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:16.094 "name": "Existed_Raid", 00:09:16.094 "uuid": "3455e8cc-fde8-44fe-a9af-1a249458b960", 00:09:16.094 "strip_size_kb": 0, 00:09:16.094 "state": "online", 00:09:16.094 "raid_level": "raid1", 00:09:16.094 "superblock": false, 00:09:16.094 "num_base_bdevs": 2, 00:09:16.094 "num_base_bdevs_discovered": 1, 00:09:16.094 "num_base_bdevs_operational": 1, 00:09:16.094 "base_bdevs_list": [ 00:09:16.094 { 00:09:16.094 "name": null, 00:09:16.094 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.094 "is_configured": false, 00:09:16.094 "data_offset": 0, 00:09:16.094 "data_size": 65536 00:09:16.094 }, 00:09:16.094 { 00:09:16.094 "name": "BaseBdev2", 00:09:16.094 "uuid": "b68d43bc-9aba-4d47-b12c-61661c0dfa03", 00:09:16.094 "is_configured": true, 00:09:16.094 "data_offset": 0, 00:09:16.094 "data_size": 65536 00:09:16.094 } 00:09:16.094 ] 00:09:16.094 }' 00:09:16.094 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:16.094 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.352 [2024-11-20 15:57:14.427632] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:16.352 [2024-11-20 15:57:14.427730] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:16.352 [2024-11-20 15:57:14.474129] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:16.352 [2024-11-20 15:57:14.474171] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:16.352 [2024-11-20 15:57:14.474180] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 61280 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 61280 ']' 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 61280 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 61280 00:09:16.352 killing process with pid 61280 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 61280' 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 61280 00:09:16.352 [2024-11-20 15:57:14.534193] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:16.352 15:57:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 61280 00:09:16.352 [2024-11-20 15:57:14.542543] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:16.917 00:09:16.917 real 0m3.417s 00:09:16.917 user 0m4.972s 00:09:16.917 sys 0m0.535s 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.917 ************************************ 00:09:16.917 END TEST raid_state_function_test 00:09:16.917 ************************************ 00:09:16.917 15:57:15 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 2 true 00:09:16.917 15:57:15 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:16.917 15:57:15 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:16.917 15:57:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:16.917 ************************************ 00:09:16.917 START TEST raid_state_function_test_sb 00:09:16.917 ************************************ 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:16.917 Process raid pid: 61517 00:09:16.917 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=61517 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 61517' 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 61517 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 61517 ']' 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:16.917 15:57:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:17.174 [2024-11-20 15:57:15.217398] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:09:17.174 [2024-11-20 15:57:15.217514] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:17.174 [2024-11-20 15:57:15.377257] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:17.431 [2024-11-20 15:57:15.476115] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:17.431 [2024-11-20 15:57:15.612105] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:17.431 [2024-11-20 15:57:15.612144] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:17.997 [2024-11-20 15:57:16.058605] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:17.997 [2024-11-20 15:57:16.058780] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:17.997 [2024-11-20 15:57:16.058853] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:17.997 [2024-11-20 15:57:16.058882] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:17.997 "name": "Existed_Raid", 00:09:17.997 "uuid": "65659312-d913-45b6-80be-82ab372e1db2", 00:09:17.997 "strip_size_kb": 0, 00:09:17.997 "state": "configuring", 00:09:17.997 "raid_level": "raid1", 00:09:17.997 "superblock": true, 00:09:17.997 "num_base_bdevs": 2, 00:09:17.997 "num_base_bdevs_discovered": 0, 00:09:17.997 "num_base_bdevs_operational": 2, 00:09:17.997 "base_bdevs_list": [ 00:09:17.997 { 00:09:17.997 "name": "BaseBdev1", 00:09:17.997 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.997 "is_configured": false, 00:09:17.997 "data_offset": 0, 00:09:17.997 "data_size": 0 00:09:17.997 }, 00:09:17.997 { 00:09:17.997 "name": "BaseBdev2", 00:09:17.997 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.997 "is_configured": false, 00:09:17.997 "data_offset": 0, 00:09:17.997 "data_size": 0 00:09:17.997 } 00:09:17.997 ] 00:09:17.997 }' 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:17.997 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.254 [2024-11-20 15:57:16.374618] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:18.254 [2024-11-20 15:57:16.374753] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.254 [2024-11-20 15:57:16.382623] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:18.254 [2024-11-20 15:57:16.382754] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:18.254 [2024-11-20 15:57:16.382769] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:18.254 [2024-11-20 15:57:16.382781] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.254 [2024-11-20 15:57:16.414894] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:18.254 BaseBdev1 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.254 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.254 [ 00:09:18.254 { 00:09:18.254 "name": "BaseBdev1", 00:09:18.254 "aliases": [ 00:09:18.254 "3cf8ed90-a069-4b72-9e08-50ef67bd40b4" 00:09:18.254 ], 00:09:18.254 "product_name": "Malloc disk", 00:09:18.254 "block_size": 512, 00:09:18.254 "num_blocks": 65536, 00:09:18.254 "uuid": "3cf8ed90-a069-4b72-9e08-50ef67bd40b4", 00:09:18.254 "assigned_rate_limits": { 00:09:18.254 "rw_ios_per_sec": 0, 00:09:18.254 "rw_mbytes_per_sec": 0, 00:09:18.254 "r_mbytes_per_sec": 0, 00:09:18.254 "w_mbytes_per_sec": 0 00:09:18.254 }, 00:09:18.254 "claimed": true, 00:09:18.254 "claim_type": "exclusive_write", 00:09:18.254 "zoned": false, 00:09:18.254 "supported_io_types": { 00:09:18.254 "read": true, 00:09:18.254 "write": true, 00:09:18.254 "unmap": true, 00:09:18.254 "flush": true, 00:09:18.254 "reset": true, 00:09:18.254 "nvme_admin": false, 00:09:18.254 "nvme_io": false, 00:09:18.254 "nvme_io_md": false, 00:09:18.254 "write_zeroes": true, 00:09:18.254 "zcopy": true, 00:09:18.254 "get_zone_info": false, 00:09:18.254 "zone_management": false, 00:09:18.254 "zone_append": false, 00:09:18.254 "compare": false, 00:09:18.254 "compare_and_write": false, 00:09:18.254 "abort": true, 00:09:18.254 "seek_hole": false, 00:09:18.254 "seek_data": false, 00:09:18.254 "copy": true, 00:09:18.254 "nvme_iov_md": false 00:09:18.255 }, 00:09:18.255 "memory_domains": [ 00:09:18.255 { 00:09:18.255 "dma_device_id": "system", 00:09:18.255 "dma_device_type": 1 00:09:18.255 }, 00:09:18.255 { 00:09:18.255 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.255 "dma_device_type": 2 00:09:18.255 } 00:09:18.255 ], 00:09:18.255 "driver_specific": {} 00:09:18.255 } 00:09:18.255 ] 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:18.255 "name": "Existed_Raid", 00:09:18.255 "uuid": "86c036b6-9a09-4869-9421-30ff2d4f5423", 00:09:18.255 "strip_size_kb": 0, 00:09:18.255 "state": "configuring", 00:09:18.255 "raid_level": "raid1", 00:09:18.255 "superblock": true, 00:09:18.255 "num_base_bdevs": 2, 00:09:18.255 "num_base_bdevs_discovered": 1, 00:09:18.255 "num_base_bdevs_operational": 2, 00:09:18.255 "base_bdevs_list": [ 00:09:18.255 { 00:09:18.255 "name": "BaseBdev1", 00:09:18.255 "uuid": "3cf8ed90-a069-4b72-9e08-50ef67bd40b4", 00:09:18.255 "is_configured": true, 00:09:18.255 "data_offset": 2048, 00:09:18.255 "data_size": 63488 00:09:18.255 }, 00:09:18.255 { 00:09:18.255 "name": "BaseBdev2", 00:09:18.255 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.255 "is_configured": false, 00:09:18.255 "data_offset": 0, 00:09:18.255 "data_size": 0 00:09:18.255 } 00:09:18.255 ] 00:09:18.255 }' 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:18.255 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.512 [2024-11-20 15:57:16.723007] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:18.512 [2024-11-20 15:57:16.723159] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.512 [2024-11-20 15:57:16.731055] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:18.512 [2024-11-20 15:57:16.732999] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:18.512 [2024-11-20 15:57:16.733116] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.512 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.768 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:18.768 "name": "Existed_Raid", 00:09:18.768 "uuid": "a17cbf63-8907-45b4-9801-d2e01abf440e", 00:09:18.768 "strip_size_kb": 0, 00:09:18.768 "state": "configuring", 00:09:18.768 "raid_level": "raid1", 00:09:18.768 "superblock": true, 00:09:18.768 "num_base_bdevs": 2, 00:09:18.768 "num_base_bdevs_discovered": 1, 00:09:18.768 "num_base_bdevs_operational": 2, 00:09:18.769 "base_bdevs_list": [ 00:09:18.769 { 00:09:18.769 "name": "BaseBdev1", 00:09:18.769 "uuid": "3cf8ed90-a069-4b72-9e08-50ef67bd40b4", 00:09:18.769 "is_configured": true, 00:09:18.769 "data_offset": 2048, 00:09:18.769 "data_size": 63488 00:09:18.769 }, 00:09:18.769 { 00:09:18.769 "name": "BaseBdev2", 00:09:18.769 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.769 "is_configured": false, 00:09:18.769 "data_offset": 0, 00:09:18.769 "data_size": 0 00:09:18.769 } 00:09:18.769 ] 00:09:18.769 }' 00:09:18.769 15:57:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:18.769 15:57:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.026 [2024-11-20 15:57:17.062175] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:19.026 BaseBdev2 00:09:19.026 [2024-11-20 15:57:17.062531] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:19.026 [2024-11-20 15:57:17.062549] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:19.026 [2024-11-20 15:57:17.062825] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:09:19.026 [2024-11-20 15:57:17.062966] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:19.026 [2024-11-20 15:57:17.062983] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:09:19.026 [2024-11-20 15:57:17.063110] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.026 [ 00:09:19.026 { 00:09:19.026 "name": "BaseBdev2", 00:09:19.026 "aliases": [ 00:09:19.026 "c0e507f7-6fb3-4725-8b58-812566e80813" 00:09:19.026 ], 00:09:19.026 "product_name": "Malloc disk", 00:09:19.026 "block_size": 512, 00:09:19.026 "num_blocks": 65536, 00:09:19.026 "uuid": "c0e507f7-6fb3-4725-8b58-812566e80813", 00:09:19.026 "assigned_rate_limits": { 00:09:19.026 "rw_ios_per_sec": 0, 00:09:19.026 "rw_mbytes_per_sec": 0, 00:09:19.026 "r_mbytes_per_sec": 0, 00:09:19.026 "w_mbytes_per_sec": 0 00:09:19.026 }, 00:09:19.026 "claimed": true, 00:09:19.026 "claim_type": "exclusive_write", 00:09:19.026 "zoned": false, 00:09:19.026 "supported_io_types": { 00:09:19.026 "read": true, 00:09:19.026 "write": true, 00:09:19.026 "unmap": true, 00:09:19.026 "flush": true, 00:09:19.026 "reset": true, 00:09:19.026 "nvme_admin": false, 00:09:19.026 "nvme_io": false, 00:09:19.026 "nvme_io_md": false, 00:09:19.026 "write_zeroes": true, 00:09:19.026 "zcopy": true, 00:09:19.026 "get_zone_info": false, 00:09:19.026 "zone_management": false, 00:09:19.026 "zone_append": false, 00:09:19.026 "compare": false, 00:09:19.026 "compare_and_write": false, 00:09:19.026 "abort": true, 00:09:19.026 "seek_hole": false, 00:09:19.026 "seek_data": false, 00:09:19.026 "copy": true, 00:09:19.026 "nvme_iov_md": false 00:09:19.026 }, 00:09:19.026 "memory_domains": [ 00:09:19.026 { 00:09:19.026 "dma_device_id": "system", 00:09:19.026 "dma_device_type": 1 00:09:19.026 }, 00:09:19.026 { 00:09:19.026 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:19.026 "dma_device_type": 2 00:09:19.026 } 00:09:19.026 ], 00:09:19.026 "driver_specific": {} 00:09:19.026 } 00:09:19.026 ] 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.026 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:19.026 "name": "Existed_Raid", 00:09:19.026 "uuid": "a17cbf63-8907-45b4-9801-d2e01abf440e", 00:09:19.026 "strip_size_kb": 0, 00:09:19.026 "state": "online", 00:09:19.027 "raid_level": "raid1", 00:09:19.027 "superblock": true, 00:09:19.027 "num_base_bdevs": 2, 00:09:19.027 "num_base_bdevs_discovered": 2, 00:09:19.027 "num_base_bdevs_operational": 2, 00:09:19.027 "base_bdevs_list": [ 00:09:19.027 { 00:09:19.027 "name": "BaseBdev1", 00:09:19.027 "uuid": "3cf8ed90-a069-4b72-9e08-50ef67bd40b4", 00:09:19.027 "is_configured": true, 00:09:19.027 "data_offset": 2048, 00:09:19.027 "data_size": 63488 00:09:19.027 }, 00:09:19.027 { 00:09:19.027 "name": "BaseBdev2", 00:09:19.027 "uuid": "c0e507f7-6fb3-4725-8b58-812566e80813", 00:09:19.027 "is_configured": true, 00:09:19.027 "data_offset": 2048, 00:09:19.027 "data_size": 63488 00:09:19.027 } 00:09:19.027 ] 00:09:19.027 }' 00:09:19.027 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:19.027 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.283 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:19.283 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:19.283 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:19.283 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:19.283 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:19.283 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:19.283 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:19.283 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.283 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.283 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:19.283 [2024-11-20 15:57:17.374584] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:19.283 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.283 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:19.283 "name": "Existed_Raid", 00:09:19.283 "aliases": [ 00:09:19.283 "a17cbf63-8907-45b4-9801-d2e01abf440e" 00:09:19.283 ], 00:09:19.283 "product_name": "Raid Volume", 00:09:19.283 "block_size": 512, 00:09:19.283 "num_blocks": 63488, 00:09:19.283 "uuid": "a17cbf63-8907-45b4-9801-d2e01abf440e", 00:09:19.283 "assigned_rate_limits": { 00:09:19.283 "rw_ios_per_sec": 0, 00:09:19.283 "rw_mbytes_per_sec": 0, 00:09:19.283 "r_mbytes_per_sec": 0, 00:09:19.283 "w_mbytes_per_sec": 0 00:09:19.283 }, 00:09:19.283 "claimed": false, 00:09:19.283 "zoned": false, 00:09:19.283 "supported_io_types": { 00:09:19.283 "read": true, 00:09:19.283 "write": true, 00:09:19.283 "unmap": false, 00:09:19.283 "flush": false, 00:09:19.283 "reset": true, 00:09:19.283 "nvme_admin": false, 00:09:19.283 "nvme_io": false, 00:09:19.283 "nvme_io_md": false, 00:09:19.283 "write_zeroes": true, 00:09:19.283 "zcopy": false, 00:09:19.283 "get_zone_info": false, 00:09:19.283 "zone_management": false, 00:09:19.283 "zone_append": false, 00:09:19.283 "compare": false, 00:09:19.283 "compare_and_write": false, 00:09:19.283 "abort": false, 00:09:19.283 "seek_hole": false, 00:09:19.283 "seek_data": false, 00:09:19.283 "copy": false, 00:09:19.283 "nvme_iov_md": false 00:09:19.283 }, 00:09:19.283 "memory_domains": [ 00:09:19.283 { 00:09:19.283 "dma_device_id": "system", 00:09:19.283 "dma_device_type": 1 00:09:19.283 }, 00:09:19.283 { 00:09:19.283 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:19.283 "dma_device_type": 2 00:09:19.283 }, 00:09:19.283 { 00:09:19.283 "dma_device_id": "system", 00:09:19.283 "dma_device_type": 1 00:09:19.283 }, 00:09:19.283 { 00:09:19.283 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:19.283 "dma_device_type": 2 00:09:19.283 } 00:09:19.283 ], 00:09:19.283 "driver_specific": { 00:09:19.283 "raid": { 00:09:19.283 "uuid": "a17cbf63-8907-45b4-9801-d2e01abf440e", 00:09:19.283 "strip_size_kb": 0, 00:09:19.283 "state": "online", 00:09:19.283 "raid_level": "raid1", 00:09:19.284 "superblock": true, 00:09:19.284 "num_base_bdevs": 2, 00:09:19.284 "num_base_bdevs_discovered": 2, 00:09:19.284 "num_base_bdevs_operational": 2, 00:09:19.284 "base_bdevs_list": [ 00:09:19.284 { 00:09:19.284 "name": "BaseBdev1", 00:09:19.284 "uuid": "3cf8ed90-a069-4b72-9e08-50ef67bd40b4", 00:09:19.284 "is_configured": true, 00:09:19.284 "data_offset": 2048, 00:09:19.284 "data_size": 63488 00:09:19.284 }, 00:09:19.284 { 00:09:19.284 "name": "BaseBdev2", 00:09:19.284 "uuid": "c0e507f7-6fb3-4725-8b58-812566e80813", 00:09:19.284 "is_configured": true, 00:09:19.284 "data_offset": 2048, 00:09:19.284 "data_size": 63488 00:09:19.284 } 00:09:19.284 ] 00:09:19.284 } 00:09:19.284 } 00:09:19.284 }' 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:19.284 BaseBdev2' 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.284 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.543 [2024-11-20 15:57:17.542372] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:19.543 "name": "Existed_Raid", 00:09:19.543 "uuid": "a17cbf63-8907-45b4-9801-d2e01abf440e", 00:09:19.543 "strip_size_kb": 0, 00:09:19.543 "state": "online", 00:09:19.543 "raid_level": "raid1", 00:09:19.543 "superblock": true, 00:09:19.543 "num_base_bdevs": 2, 00:09:19.543 "num_base_bdevs_discovered": 1, 00:09:19.543 "num_base_bdevs_operational": 1, 00:09:19.543 "base_bdevs_list": [ 00:09:19.543 { 00:09:19.543 "name": null, 00:09:19.543 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:19.543 "is_configured": false, 00:09:19.543 "data_offset": 0, 00:09:19.543 "data_size": 63488 00:09:19.543 }, 00:09:19.543 { 00:09:19.543 "name": "BaseBdev2", 00:09:19.543 "uuid": "c0e507f7-6fb3-4725-8b58-812566e80813", 00:09:19.543 "is_configured": true, 00:09:19.543 "data_offset": 2048, 00:09:19.543 "data_size": 63488 00:09:19.543 } 00:09:19.543 ] 00:09:19.543 }' 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:19.543 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.800 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:19.800 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:19.800 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:19.800 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.800 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.800 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.800 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.800 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:19.800 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:19.800 15:57:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:19.800 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.800 15:57:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.800 [2024-11-20 15:57:17.961634] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:19.800 [2024-11-20 15:57:17.961837] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:19.800 [2024-11-20 15:57:18.021044] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:19.800 [2024-11-20 15:57:18.021224] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:19.800 [2024-11-20 15:57:18.021304] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:09:19.800 15:57:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.800 15:57:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:19.800 15:57:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:19.800 15:57:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:19.800 15:57:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.800 15:57:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.800 15:57:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.800 15:57:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.057 15:57:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:20.057 15:57:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:20.057 15:57:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:09:20.057 15:57:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 61517 00:09:20.057 15:57:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 61517 ']' 00:09:20.057 15:57:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 61517 00:09:20.057 15:57:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:09:20.057 15:57:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:20.057 15:57:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 61517 00:09:20.057 killing process with pid 61517 00:09:20.057 15:57:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:20.057 15:57:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:20.057 15:57:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 61517' 00:09:20.057 15:57:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 61517 00:09:20.057 [2024-11-20 15:57:18.075280] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:20.057 15:57:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 61517 00:09:20.057 [2024-11-20 15:57:18.085842] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:20.621 15:57:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:20.621 00:09:20.621 real 0m3.648s 00:09:20.621 user 0m5.237s 00:09:20.621 sys 0m0.548s 00:09:20.621 15:57:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:20.621 15:57:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.621 ************************************ 00:09:20.621 END TEST raid_state_function_test_sb 00:09:20.621 ************************************ 00:09:20.621 15:57:18 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 2 00:09:20.621 15:57:18 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:09:20.621 15:57:18 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:20.621 15:57:18 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:20.621 ************************************ 00:09:20.621 START TEST raid_superblock_test 00:09:20.621 ************************************ 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=61753 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 61753 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 61753 ']' 00:09:20.621 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:20.621 15:57:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.877 [2024-11-20 15:57:18.933223] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:09:20.877 [2024-11-20 15:57:18.933581] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid61753 ] 00:09:20.877 [2024-11-20 15:57:19.118096] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:21.133 [2024-11-20 15:57:19.250182] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:21.389 [2024-11-20 15:57:19.415562] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:21.389 [2024-11-20 15:57:19.415612] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.646 malloc1 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.646 [2024-11-20 15:57:19.879977] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:21.646 [2024-11-20 15:57:19.880150] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:21.646 [2024-11-20 15:57:19.880193] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:21.646 [2024-11-20 15:57:19.880776] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:21.646 [2024-11-20 15:57:19.883086] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:21.646 [2024-11-20 15:57:19.883207] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:21.646 pt1 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.646 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.931 malloc2 00:09:21.931 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.931 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:21.931 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.931 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.931 [2024-11-20 15:57:19.924551] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:21.931 [2024-11-20 15:57:19.924704] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:21.931 [2024-11-20 15:57:19.924734] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:21.931 [2024-11-20 15:57:19.924744] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:21.931 [2024-11-20 15:57:19.926861] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:21.931 [2024-11-20 15:57:19.926893] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:21.931 pt2 00:09:21.931 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.931 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:21.931 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:21.931 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:09:21.931 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.931 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.931 [2024-11-20 15:57:19.932606] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:21.932 [2024-11-20 15:57:19.934523] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:21.932 [2024-11-20 15:57:19.934764] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:09:21.932 [2024-11-20 15:57:19.934843] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:21.932 [2024-11-20 15:57:19.935110] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:09:21.932 [2024-11-20 15:57:19.935313] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:09:21.932 [2024-11-20 15:57:19.935390] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:09:21.932 [2024-11-20 15:57:19.935596] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:21.932 "name": "raid_bdev1", 00:09:21.932 "uuid": "e355767c-c37a-40c8-8a63-4df2088884e9", 00:09:21.932 "strip_size_kb": 0, 00:09:21.932 "state": "online", 00:09:21.932 "raid_level": "raid1", 00:09:21.932 "superblock": true, 00:09:21.932 "num_base_bdevs": 2, 00:09:21.932 "num_base_bdevs_discovered": 2, 00:09:21.932 "num_base_bdevs_operational": 2, 00:09:21.932 "base_bdevs_list": [ 00:09:21.932 { 00:09:21.932 "name": "pt1", 00:09:21.932 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:21.932 "is_configured": true, 00:09:21.932 "data_offset": 2048, 00:09:21.932 "data_size": 63488 00:09:21.932 }, 00:09:21.932 { 00:09:21.932 "name": "pt2", 00:09:21.932 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:21.932 "is_configured": true, 00:09:21.932 "data_offset": 2048, 00:09:21.932 "data_size": 63488 00:09:21.932 } 00:09:21.932 ] 00:09:21.932 }' 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:21.932 15:57:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.201 [2024-11-20 15:57:20.252967] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:22.201 "name": "raid_bdev1", 00:09:22.201 "aliases": [ 00:09:22.201 "e355767c-c37a-40c8-8a63-4df2088884e9" 00:09:22.201 ], 00:09:22.201 "product_name": "Raid Volume", 00:09:22.201 "block_size": 512, 00:09:22.201 "num_blocks": 63488, 00:09:22.201 "uuid": "e355767c-c37a-40c8-8a63-4df2088884e9", 00:09:22.201 "assigned_rate_limits": { 00:09:22.201 "rw_ios_per_sec": 0, 00:09:22.201 "rw_mbytes_per_sec": 0, 00:09:22.201 "r_mbytes_per_sec": 0, 00:09:22.201 "w_mbytes_per_sec": 0 00:09:22.201 }, 00:09:22.201 "claimed": false, 00:09:22.201 "zoned": false, 00:09:22.201 "supported_io_types": { 00:09:22.201 "read": true, 00:09:22.201 "write": true, 00:09:22.201 "unmap": false, 00:09:22.201 "flush": false, 00:09:22.201 "reset": true, 00:09:22.201 "nvme_admin": false, 00:09:22.201 "nvme_io": false, 00:09:22.201 "nvme_io_md": false, 00:09:22.201 "write_zeroes": true, 00:09:22.201 "zcopy": false, 00:09:22.201 "get_zone_info": false, 00:09:22.201 "zone_management": false, 00:09:22.201 "zone_append": false, 00:09:22.201 "compare": false, 00:09:22.201 "compare_and_write": false, 00:09:22.201 "abort": false, 00:09:22.201 "seek_hole": false, 00:09:22.201 "seek_data": false, 00:09:22.201 "copy": false, 00:09:22.201 "nvme_iov_md": false 00:09:22.201 }, 00:09:22.201 "memory_domains": [ 00:09:22.201 { 00:09:22.201 "dma_device_id": "system", 00:09:22.201 "dma_device_type": 1 00:09:22.201 }, 00:09:22.201 { 00:09:22.201 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:22.201 "dma_device_type": 2 00:09:22.201 }, 00:09:22.201 { 00:09:22.201 "dma_device_id": "system", 00:09:22.201 "dma_device_type": 1 00:09:22.201 }, 00:09:22.201 { 00:09:22.201 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:22.201 "dma_device_type": 2 00:09:22.201 } 00:09:22.201 ], 00:09:22.201 "driver_specific": { 00:09:22.201 "raid": { 00:09:22.201 "uuid": "e355767c-c37a-40c8-8a63-4df2088884e9", 00:09:22.201 "strip_size_kb": 0, 00:09:22.201 "state": "online", 00:09:22.201 "raid_level": "raid1", 00:09:22.201 "superblock": true, 00:09:22.201 "num_base_bdevs": 2, 00:09:22.201 "num_base_bdevs_discovered": 2, 00:09:22.201 "num_base_bdevs_operational": 2, 00:09:22.201 "base_bdevs_list": [ 00:09:22.201 { 00:09:22.201 "name": "pt1", 00:09:22.201 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:22.201 "is_configured": true, 00:09:22.201 "data_offset": 2048, 00:09:22.201 "data_size": 63488 00:09:22.201 }, 00:09:22.201 { 00:09:22.201 "name": "pt2", 00:09:22.201 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:22.201 "is_configured": true, 00:09:22.201 "data_offset": 2048, 00:09:22.201 "data_size": 63488 00:09:22.201 } 00:09:22.201 ] 00:09:22.201 } 00:09:22.201 } 00:09:22.201 }' 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:22.201 pt2' 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:22.201 [2024-11-20 15:57:20.412990] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=e355767c-c37a-40c8-8a63-4df2088884e9 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z e355767c-c37a-40c8-8a63-4df2088884e9 ']' 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.201 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.201 [2024-11-20 15:57:20.444700] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:22.201 [2024-11-20 15:57:20.444798] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:22.202 [2024-11-20 15:57:20.444915] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:22.202 [2024-11-20 15:57:20.444992] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:22.202 [2024-11-20 15:57:20.445469] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:09:22.202 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:09:22.459 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.460 [2024-11-20 15:57:20.548747] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:22.460 [2024-11-20 15:57:20.550757] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:22.460 [2024-11-20 15:57:20.550908] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:22.460 [2024-11-20 15:57:20.550963] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:22.460 [2024-11-20 15:57:20.550978] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:22.460 [2024-11-20 15:57:20.550988] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:09:22.460 request: 00:09:22.460 { 00:09:22.460 "name": "raid_bdev1", 00:09:22.460 "raid_level": "raid1", 00:09:22.460 "base_bdevs": [ 00:09:22.460 "malloc1", 00:09:22.460 "malloc2" 00:09:22.460 ], 00:09:22.460 "superblock": false, 00:09:22.460 "method": "bdev_raid_create", 00:09:22.460 "req_id": 1 00:09:22.460 } 00:09:22.460 Got JSON-RPC error response 00:09:22.460 response: 00:09:22.460 { 00:09:22.460 "code": -17, 00:09:22.460 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:22.460 } 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.460 [2024-11-20 15:57:20.592738] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:22.460 [2024-11-20 15:57:20.592871] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:22.460 [2024-11-20 15:57:20.592937] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:22.460 [2024-11-20 15:57:20.593046] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:22.460 [2024-11-20 15:57:20.595292] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:22.460 [2024-11-20 15:57:20.595406] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:22.460 pt1 00:09:22.460 [2024-11-20 15:57:20.595537] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:22.460 [2024-11-20 15:57:20.595590] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:22.460 "name": "raid_bdev1", 00:09:22.460 "uuid": "e355767c-c37a-40c8-8a63-4df2088884e9", 00:09:22.460 "strip_size_kb": 0, 00:09:22.460 "state": "configuring", 00:09:22.460 "raid_level": "raid1", 00:09:22.460 "superblock": true, 00:09:22.460 "num_base_bdevs": 2, 00:09:22.460 "num_base_bdevs_discovered": 1, 00:09:22.460 "num_base_bdevs_operational": 2, 00:09:22.460 "base_bdevs_list": [ 00:09:22.460 { 00:09:22.460 "name": "pt1", 00:09:22.460 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:22.460 "is_configured": true, 00:09:22.460 "data_offset": 2048, 00:09:22.460 "data_size": 63488 00:09:22.460 }, 00:09:22.460 { 00:09:22.460 "name": null, 00:09:22.460 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:22.460 "is_configured": false, 00:09:22.460 "data_offset": 2048, 00:09:22.460 "data_size": 63488 00:09:22.460 } 00:09:22.460 ] 00:09:22.460 }' 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:22.460 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.717 [2024-11-20 15:57:20.908824] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:22.717 [2024-11-20 15:57:20.908988] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:22.717 [2024-11-20 15:57:20.909026] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:09:22.717 [2024-11-20 15:57:20.909086] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:22.717 [2024-11-20 15:57:20.909518] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:22.717 [2024-11-20 15:57:20.909624] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:22.717 [2024-11-20 15:57:20.909759] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:22.717 [2024-11-20 15:57:20.909852] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:22.717 [2024-11-20 15:57:20.909972] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:22.717 [2024-11-20 15:57:20.909984] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:22.717 [2024-11-20 15:57:20.910223] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:09:22.717 [2024-11-20 15:57:20.910351] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:22.717 [2024-11-20 15:57:20.910359] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:09:22.717 [2024-11-20 15:57:20.910484] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:22.717 pt2 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.717 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:22.717 "name": "raid_bdev1", 00:09:22.717 "uuid": "e355767c-c37a-40c8-8a63-4df2088884e9", 00:09:22.718 "strip_size_kb": 0, 00:09:22.718 "state": "online", 00:09:22.718 "raid_level": "raid1", 00:09:22.718 "superblock": true, 00:09:22.718 "num_base_bdevs": 2, 00:09:22.718 "num_base_bdevs_discovered": 2, 00:09:22.718 "num_base_bdevs_operational": 2, 00:09:22.718 "base_bdevs_list": [ 00:09:22.718 { 00:09:22.718 "name": "pt1", 00:09:22.718 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:22.718 "is_configured": true, 00:09:22.718 "data_offset": 2048, 00:09:22.718 "data_size": 63488 00:09:22.718 }, 00:09:22.718 { 00:09:22.718 "name": "pt2", 00:09:22.718 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:22.718 "is_configured": true, 00:09:22.718 "data_offset": 2048, 00:09:22.718 "data_size": 63488 00:09:22.718 } 00:09:22.718 ] 00:09:22.718 }' 00:09:22.718 15:57:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:22.718 15:57:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.281 [2024-11-20 15:57:21.245151] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:23.281 "name": "raid_bdev1", 00:09:23.281 "aliases": [ 00:09:23.281 "e355767c-c37a-40c8-8a63-4df2088884e9" 00:09:23.281 ], 00:09:23.281 "product_name": "Raid Volume", 00:09:23.281 "block_size": 512, 00:09:23.281 "num_blocks": 63488, 00:09:23.281 "uuid": "e355767c-c37a-40c8-8a63-4df2088884e9", 00:09:23.281 "assigned_rate_limits": { 00:09:23.281 "rw_ios_per_sec": 0, 00:09:23.281 "rw_mbytes_per_sec": 0, 00:09:23.281 "r_mbytes_per_sec": 0, 00:09:23.281 "w_mbytes_per_sec": 0 00:09:23.281 }, 00:09:23.281 "claimed": false, 00:09:23.281 "zoned": false, 00:09:23.281 "supported_io_types": { 00:09:23.281 "read": true, 00:09:23.281 "write": true, 00:09:23.281 "unmap": false, 00:09:23.281 "flush": false, 00:09:23.281 "reset": true, 00:09:23.281 "nvme_admin": false, 00:09:23.281 "nvme_io": false, 00:09:23.281 "nvme_io_md": false, 00:09:23.281 "write_zeroes": true, 00:09:23.281 "zcopy": false, 00:09:23.281 "get_zone_info": false, 00:09:23.281 "zone_management": false, 00:09:23.281 "zone_append": false, 00:09:23.281 "compare": false, 00:09:23.281 "compare_and_write": false, 00:09:23.281 "abort": false, 00:09:23.281 "seek_hole": false, 00:09:23.281 "seek_data": false, 00:09:23.281 "copy": false, 00:09:23.281 "nvme_iov_md": false 00:09:23.281 }, 00:09:23.281 "memory_domains": [ 00:09:23.281 { 00:09:23.281 "dma_device_id": "system", 00:09:23.281 "dma_device_type": 1 00:09:23.281 }, 00:09:23.281 { 00:09:23.281 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:23.281 "dma_device_type": 2 00:09:23.281 }, 00:09:23.281 { 00:09:23.281 "dma_device_id": "system", 00:09:23.281 "dma_device_type": 1 00:09:23.281 }, 00:09:23.281 { 00:09:23.281 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:23.281 "dma_device_type": 2 00:09:23.281 } 00:09:23.281 ], 00:09:23.281 "driver_specific": { 00:09:23.281 "raid": { 00:09:23.281 "uuid": "e355767c-c37a-40c8-8a63-4df2088884e9", 00:09:23.281 "strip_size_kb": 0, 00:09:23.281 "state": "online", 00:09:23.281 "raid_level": "raid1", 00:09:23.281 "superblock": true, 00:09:23.281 "num_base_bdevs": 2, 00:09:23.281 "num_base_bdevs_discovered": 2, 00:09:23.281 "num_base_bdevs_operational": 2, 00:09:23.281 "base_bdevs_list": [ 00:09:23.281 { 00:09:23.281 "name": "pt1", 00:09:23.281 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:23.281 "is_configured": true, 00:09:23.281 "data_offset": 2048, 00:09:23.281 "data_size": 63488 00:09:23.281 }, 00:09:23.281 { 00:09:23.281 "name": "pt2", 00:09:23.281 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:23.281 "is_configured": true, 00:09:23.281 "data_offset": 2048, 00:09:23.281 "data_size": 63488 00:09:23.281 } 00:09:23.281 ] 00:09:23.281 } 00:09:23.281 } 00:09:23.281 }' 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:23.281 pt2' 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.281 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.282 [2024-11-20 15:57:21.405175] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' e355767c-c37a-40c8-8a63-4df2088884e9 '!=' e355767c-c37a-40c8-8a63-4df2088884e9 ']' 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.282 [2024-11-20 15:57:21.428959] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:23.282 "name": "raid_bdev1", 00:09:23.282 "uuid": "e355767c-c37a-40c8-8a63-4df2088884e9", 00:09:23.282 "strip_size_kb": 0, 00:09:23.282 "state": "online", 00:09:23.282 "raid_level": "raid1", 00:09:23.282 "superblock": true, 00:09:23.282 "num_base_bdevs": 2, 00:09:23.282 "num_base_bdevs_discovered": 1, 00:09:23.282 "num_base_bdevs_operational": 1, 00:09:23.282 "base_bdevs_list": [ 00:09:23.282 { 00:09:23.282 "name": null, 00:09:23.282 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:23.282 "is_configured": false, 00:09:23.282 "data_offset": 0, 00:09:23.282 "data_size": 63488 00:09:23.282 }, 00:09:23.282 { 00:09:23.282 "name": "pt2", 00:09:23.282 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:23.282 "is_configured": true, 00:09:23.282 "data_offset": 2048, 00:09:23.282 "data_size": 63488 00:09:23.282 } 00:09:23.282 ] 00:09:23.282 }' 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:23.282 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.539 [2024-11-20 15:57:21.733003] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:23.539 [2024-11-20 15:57:21.733119] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:23.539 [2024-11-20 15:57:21.733191] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:23.539 [2024-11-20 15:57:21.733235] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:23.539 [2024-11-20 15:57:21.733247] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=1 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.539 [2024-11-20 15:57:21.781000] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:23.539 [2024-11-20 15:57:21.781050] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:23.539 [2024-11-20 15:57:21.781063] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:23.539 [2024-11-20 15:57:21.781074] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:23.539 [2024-11-20 15:57:21.783277] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:23.539 [2024-11-20 15:57:21.783314] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:23.539 [2024-11-20 15:57:21.783383] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:23.539 [2024-11-20 15:57:21.783423] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:23.539 [2024-11-20 15:57:21.783511] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:09:23.539 [2024-11-20 15:57:21.783523] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:23.539 [2024-11-20 15:57:21.783766] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:09:23.539 [2024-11-20 15:57:21.783997] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:09:23.539 [2024-11-20 15:57:21.784010] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:09:23.539 [2024-11-20 15:57:21.784141] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:23.539 pt2 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:23.539 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:23.795 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:23.795 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.795 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:23.795 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.795 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.795 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:23.795 "name": "raid_bdev1", 00:09:23.795 "uuid": "e355767c-c37a-40c8-8a63-4df2088884e9", 00:09:23.795 "strip_size_kb": 0, 00:09:23.795 "state": "online", 00:09:23.795 "raid_level": "raid1", 00:09:23.795 "superblock": true, 00:09:23.795 "num_base_bdevs": 2, 00:09:23.795 "num_base_bdevs_discovered": 1, 00:09:23.795 "num_base_bdevs_operational": 1, 00:09:23.795 "base_bdevs_list": [ 00:09:23.795 { 00:09:23.795 "name": null, 00:09:23.796 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:23.796 "is_configured": false, 00:09:23.796 "data_offset": 2048, 00:09:23.796 "data_size": 63488 00:09:23.796 }, 00:09:23.796 { 00:09:23.796 "name": "pt2", 00:09:23.796 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:23.796 "is_configured": true, 00:09:23.796 "data_offset": 2048, 00:09:23.796 "data_size": 63488 00:09:23.796 } 00:09:23.796 ] 00:09:23.796 }' 00:09:23.796 15:57:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:23.796 15:57:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.053 [2024-11-20 15:57:22.105085] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:24.053 [2024-11-20 15:57:22.105221] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:24.053 [2024-11-20 15:57:22.105292] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:24.053 [2024-11-20 15:57:22.105339] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:24.053 [2024-11-20 15:57:22.105348] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.053 [2024-11-20 15:57:22.149103] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:24.053 [2024-11-20 15:57:22.149249] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:24.053 [2024-11-20 15:57:22.149289] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:09:24.053 [2024-11-20 15:57:22.149353] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:24.053 [2024-11-20 15:57:22.151604] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:24.053 [2024-11-20 15:57:22.151736] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:24.053 [2024-11-20 15:57:22.151871] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:24.053 [2024-11-20 15:57:22.151966] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:24.053 pt1 00:09:24.053 [2024-11-20 15:57:22.152148] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:09:24.053 [2024-11-20 15:57:22.152162] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:24.053 [2024-11-20 15:57:22.152178] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:09:24.053 [2024-11-20 15:57:22.152223] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:24.053 [2024-11-20 15:57:22.152290] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:09:24.053 [2024-11-20 15:57:22.152299] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:24.053 [2024-11-20 15:57:22.152548] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:09:24.053 [2024-11-20 15:57:22.152688] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:09:24.053 [2024-11-20 15:57:22.152700] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:09:24.053 [2024-11-20 15:57:22.152829] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:24.053 "name": "raid_bdev1", 00:09:24.053 "uuid": "e355767c-c37a-40c8-8a63-4df2088884e9", 00:09:24.053 "strip_size_kb": 0, 00:09:24.053 "state": "online", 00:09:24.053 "raid_level": "raid1", 00:09:24.053 "superblock": true, 00:09:24.053 "num_base_bdevs": 2, 00:09:24.053 "num_base_bdevs_discovered": 1, 00:09:24.053 "num_base_bdevs_operational": 1, 00:09:24.053 "base_bdevs_list": [ 00:09:24.053 { 00:09:24.053 "name": null, 00:09:24.053 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:24.053 "is_configured": false, 00:09:24.053 "data_offset": 2048, 00:09:24.053 "data_size": 63488 00:09:24.053 }, 00:09:24.053 { 00:09:24.053 "name": "pt2", 00:09:24.053 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:24.053 "is_configured": true, 00:09:24.053 "data_offset": 2048, 00:09:24.053 "data_size": 63488 00:09:24.053 } 00:09:24.053 ] 00:09:24.053 }' 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:24.053 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.310 [2024-11-20 15:57:22.501407] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' e355767c-c37a-40c8-8a63-4df2088884e9 '!=' e355767c-c37a-40c8-8a63-4df2088884e9 ']' 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 61753 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 61753 ']' 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 61753 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:24.310 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 61753 00:09:24.310 killing process with pid 61753 00:09:24.311 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:24.311 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:24.311 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 61753' 00:09:24.311 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 61753 00:09:24.311 [2024-11-20 15:57:22.556684] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:24.311 [2024-11-20 15:57:22.556764] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:24.311 15:57:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 61753 00:09:24.311 [2024-11-20 15:57:22.556807] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:24.311 [2024-11-20 15:57:22.556825] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:09:24.567 [2024-11-20 15:57:22.685208] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:25.497 15:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:25.497 00:09:25.497 real 0m4.553s 00:09:25.497 user 0m6.912s 00:09:25.497 sys 0m0.722s 00:09:25.497 15:57:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:25.497 15:57:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.497 ************************************ 00:09:25.497 END TEST raid_superblock_test 00:09:25.497 ************************************ 00:09:25.497 15:57:23 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 2 read 00:09:25.497 15:57:23 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:25.497 15:57:23 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:25.497 15:57:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:25.497 ************************************ 00:09:25.497 START TEST raid_read_error_test 00:09:25.497 ************************************ 00:09:25.497 15:57:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 2 read 00:09:25.497 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:25.497 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.dC1v6qIzzA 00:09:25.498 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=62072 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 62072 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 62072 ']' 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:25.498 15:57:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.498 [2024-11-20 15:57:23.503293] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:09:25.498 [2024-11-20 15:57:23.503414] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid62072 ] 00:09:25.498 [2024-11-20 15:57:23.663496] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:25.755 [2024-11-20 15:57:23.766162] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:25.755 [2024-11-20 15:57:23.901386] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:25.755 [2024-11-20 15:57:23.901430] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.319 BaseBdev1_malloc 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.319 true 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.319 [2024-11-20 15:57:24.398797] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:26.319 [2024-11-20 15:57:24.398848] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:26.319 [2024-11-20 15:57:24.398866] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:26.319 [2024-11-20 15:57:24.398877] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:26.319 [2024-11-20 15:57:24.400963] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:26.319 [2024-11-20 15:57:24.400999] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:26.319 BaseBdev1 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.319 BaseBdev2_malloc 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.319 true 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.319 [2024-11-20 15:57:24.442798] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:26.319 [2024-11-20 15:57:24.442841] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:26.319 [2024-11-20 15:57:24.442856] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:26.319 [2024-11-20 15:57:24.442866] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:26.319 [2024-11-20 15:57:24.444915] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:26.319 [2024-11-20 15:57:24.444947] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:26.319 BaseBdev2 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.319 [2024-11-20 15:57:24.450854] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:26.319 [2024-11-20 15:57:24.452662] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:26.319 [2024-11-20 15:57:24.452852] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:26.319 [2024-11-20 15:57:24.452865] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:26.319 [2024-11-20 15:57:24.453099] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:09:26.319 [2024-11-20 15:57:24.453246] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:26.319 [2024-11-20 15:57:24.453255] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:09:26.319 [2024-11-20 15:57:24.453397] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:26.319 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:26.320 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:26.320 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:26.320 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:26.320 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:26.320 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:26.320 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:26.320 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:26.320 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:26.320 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.320 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:26.320 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.320 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.320 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.320 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:26.320 "name": "raid_bdev1", 00:09:26.320 "uuid": "7aaa68c2-2a55-4f35-9ce5-a6b76145bf58", 00:09:26.320 "strip_size_kb": 0, 00:09:26.320 "state": "online", 00:09:26.320 "raid_level": "raid1", 00:09:26.320 "superblock": true, 00:09:26.320 "num_base_bdevs": 2, 00:09:26.320 "num_base_bdevs_discovered": 2, 00:09:26.320 "num_base_bdevs_operational": 2, 00:09:26.320 "base_bdevs_list": [ 00:09:26.320 { 00:09:26.320 "name": "BaseBdev1", 00:09:26.320 "uuid": "f17b2269-a0f9-5729-b2a9-6971a65d32e2", 00:09:26.320 "is_configured": true, 00:09:26.320 "data_offset": 2048, 00:09:26.320 "data_size": 63488 00:09:26.320 }, 00:09:26.320 { 00:09:26.320 "name": "BaseBdev2", 00:09:26.320 "uuid": "af4ece33-27fe-5787-8a40-77fbd14db65d", 00:09:26.320 "is_configured": true, 00:09:26.320 "data_offset": 2048, 00:09:26.320 "data_size": 63488 00:09:26.320 } 00:09:26.320 ] 00:09:26.320 }' 00:09:26.320 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:26.320 15:57:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.577 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:26.577 15:57:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:26.834 [2024-11-20 15:57:24.827878] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:27.764 "name": "raid_bdev1", 00:09:27.764 "uuid": "7aaa68c2-2a55-4f35-9ce5-a6b76145bf58", 00:09:27.764 "strip_size_kb": 0, 00:09:27.764 "state": "online", 00:09:27.764 "raid_level": "raid1", 00:09:27.764 "superblock": true, 00:09:27.764 "num_base_bdevs": 2, 00:09:27.764 "num_base_bdevs_discovered": 2, 00:09:27.764 "num_base_bdevs_operational": 2, 00:09:27.764 "base_bdevs_list": [ 00:09:27.764 { 00:09:27.764 "name": "BaseBdev1", 00:09:27.764 "uuid": "f17b2269-a0f9-5729-b2a9-6971a65d32e2", 00:09:27.764 "is_configured": true, 00:09:27.764 "data_offset": 2048, 00:09:27.764 "data_size": 63488 00:09:27.764 }, 00:09:27.764 { 00:09:27.764 "name": "BaseBdev2", 00:09:27.764 "uuid": "af4ece33-27fe-5787-8a40-77fbd14db65d", 00:09:27.764 "is_configured": true, 00:09:27.764 "data_offset": 2048, 00:09:27.764 "data_size": 63488 00:09:27.764 } 00:09:27.764 ] 00:09:27.764 }' 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:27.764 15:57:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.021 15:57:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:28.021 15:57:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.021 15:57:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.021 [2024-11-20 15:57:26.073621] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:28.021 [2024-11-20 15:57:26.073657] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:28.021 [2024-11-20 15:57:26.076726] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:28.021 [2024-11-20 15:57:26.076769] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:28.021 [2024-11-20 15:57:26.076852] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:28.021 [2024-11-20 15:57:26.076865] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:09:28.021 { 00:09:28.021 "results": [ 00:09:28.021 { 00:09:28.021 "job": "raid_bdev1", 00:09:28.021 "core_mask": "0x1", 00:09:28.021 "workload": "randrw", 00:09:28.021 "percentage": 50, 00:09:28.021 "status": "finished", 00:09:28.021 "queue_depth": 1, 00:09:28.021 "io_size": 131072, 00:09:28.021 "runtime": 1.24383, 00:09:28.021 "iops": 18003.26411165513, 00:09:28.021 "mibps": 2250.4080139568914, 00:09:28.021 "io_failed": 0, 00:09:28.021 "io_timeout": 0, 00:09:28.021 "avg_latency_us": 52.33694870306311, 00:09:28.021 "min_latency_us": 29.735384615384614, 00:09:28.022 "max_latency_us": 1676.2092307692308 00:09:28.022 } 00:09:28.022 ], 00:09:28.022 "core_count": 1 00:09:28.022 } 00:09:28.022 15:57:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.022 15:57:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 62072 00:09:28.022 15:57:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 62072 ']' 00:09:28.022 15:57:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 62072 00:09:28.022 15:57:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:09:28.022 15:57:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:28.022 15:57:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 62072 00:09:28.022 killing process with pid 62072 00:09:28.022 15:57:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:28.022 15:57:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:28.022 15:57:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 62072' 00:09:28.022 15:57:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 62072 00:09:28.022 [2024-11-20 15:57:26.104243] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:28.022 15:57:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 62072 00:09:28.022 [2024-11-20 15:57:26.189265] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:28.954 15:57:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:28.954 15:57:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.dC1v6qIzzA 00:09:28.954 15:57:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:28.954 15:57:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:28.954 15:57:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:28.954 15:57:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:28.954 15:57:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:28.954 15:57:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:28.954 00:09:28.954 real 0m3.529s 00:09:28.954 user 0m4.195s 00:09:28.954 sys 0m0.376s 00:09:28.954 ************************************ 00:09:28.954 END TEST raid_read_error_test 00:09:28.954 ************************************ 00:09:28.954 15:57:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:28.954 15:57:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.954 15:57:26 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 2 write 00:09:28.954 15:57:26 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:28.954 15:57:26 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:28.954 15:57:26 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:28.954 ************************************ 00:09:28.955 START TEST raid_write_error_test 00:09:28.955 ************************************ 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 2 write 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.XmEHGKIEV6 00:09:28.955 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=62201 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 62201 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 62201 ']' 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.955 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:28.955 [2024-11-20 15:57:27.067919] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:09:28.955 [2024-11-20 15:57:27.068019] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid62201 ] 00:09:29.212 [2024-11-20 15:57:27.224130] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:29.212 [2024-11-20 15:57:27.375971] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:29.471 [2024-11-20 15:57:27.512206] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:29.471 [2024-11-20 15:57:27.512239] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:29.730 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:29.730 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:29.730 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:29.730 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:29.730 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.730 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.730 BaseBdev1_malloc 00:09:29.730 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.730 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:29.730 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.730 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.730 true 00:09:29.730 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.730 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:29.730 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.730 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.730 [2024-11-20 15:57:27.922281] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:29.730 [2024-11-20 15:57:27.922460] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:29.730 [2024-11-20 15:57:27.922486] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:29.730 [2024-11-20 15:57:27.922498] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:29.730 [2024-11-20 15:57:27.924625] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:29.730 [2024-11-20 15:57:27.924663] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:29.730 BaseBdev1 00:09:29.730 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.730 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:29.730 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:29.731 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.731 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.731 BaseBdev2_malloc 00:09:29.731 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.731 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:29.731 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.731 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.731 true 00:09:29.731 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.731 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:29.731 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.731 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.731 [2024-11-20 15:57:27.966463] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:29.731 [2024-11-20 15:57:27.966515] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:29.731 [2024-11-20 15:57:27.966532] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:29.731 [2024-11-20 15:57:27.966543] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:29.731 [2024-11-20 15:57:27.968656] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:29.731 [2024-11-20 15:57:27.968702] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:29.731 BaseBdev2 00:09:29.731 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.731 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:09:29.731 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.731 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.731 [2024-11-20 15:57:27.974510] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:29.731 [2024-11-20 15:57:27.976346] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:29.731 [2024-11-20 15:57:27.976631] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:29.731 [2024-11-20 15:57:27.976648] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:29.731 [2024-11-20 15:57:27.976909] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:09:29.731 [2024-11-20 15:57:27.977060] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:29.731 [2024-11-20 15:57:27.977069] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:09:29.731 [2024-11-20 15:57:27.977204] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:29.990 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.990 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:29.990 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:29.990 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:29.990 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:29.990 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:29.990 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:29.990 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.990 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.990 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.990 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.990 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.990 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.990 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.990 15:57:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:29.990 15:57:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.990 15:57:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.990 "name": "raid_bdev1", 00:09:29.990 "uuid": "e236c6ba-7c47-4ec6-9234-a52b6d8e8e24", 00:09:29.990 "strip_size_kb": 0, 00:09:29.990 "state": "online", 00:09:29.990 "raid_level": "raid1", 00:09:29.990 "superblock": true, 00:09:29.990 "num_base_bdevs": 2, 00:09:29.990 "num_base_bdevs_discovered": 2, 00:09:29.990 "num_base_bdevs_operational": 2, 00:09:29.990 "base_bdevs_list": [ 00:09:29.990 { 00:09:29.990 "name": "BaseBdev1", 00:09:29.990 "uuid": "5ecbfa3a-14b6-55eb-a490-328f147e733c", 00:09:29.990 "is_configured": true, 00:09:29.990 "data_offset": 2048, 00:09:29.990 "data_size": 63488 00:09:29.990 }, 00:09:29.990 { 00:09:29.990 "name": "BaseBdev2", 00:09:29.990 "uuid": "6cd85b6a-5da5-5518-acab-deddb9d2dc19", 00:09:29.990 "is_configured": true, 00:09:29.990 "data_offset": 2048, 00:09:29.990 "data_size": 63488 00:09:29.990 } 00:09:29.990 ] 00:09:29.990 }' 00:09:29.990 15:57:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.990 15:57:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.248 15:57:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:30.248 15:57:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:30.248 [2024-11-20 15:57:28.375580] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.183 [2024-11-20 15:57:29.289049] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:09:31.183 [2024-11-20 15:57:29.289103] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:31.183 [2024-11-20 15:57:29.289283] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000063c0 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=1 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:31.183 "name": "raid_bdev1", 00:09:31.183 "uuid": "e236c6ba-7c47-4ec6-9234-a52b6d8e8e24", 00:09:31.183 "strip_size_kb": 0, 00:09:31.183 "state": "online", 00:09:31.183 "raid_level": "raid1", 00:09:31.183 "superblock": true, 00:09:31.183 "num_base_bdevs": 2, 00:09:31.183 "num_base_bdevs_discovered": 1, 00:09:31.183 "num_base_bdevs_operational": 1, 00:09:31.183 "base_bdevs_list": [ 00:09:31.183 { 00:09:31.183 "name": null, 00:09:31.183 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.183 "is_configured": false, 00:09:31.183 "data_offset": 0, 00:09:31.183 "data_size": 63488 00:09:31.183 }, 00:09:31.183 { 00:09:31.183 "name": "BaseBdev2", 00:09:31.183 "uuid": "6cd85b6a-5da5-5518-acab-deddb9d2dc19", 00:09:31.183 "is_configured": true, 00:09:31.183 "data_offset": 2048, 00:09:31.183 "data_size": 63488 00:09:31.183 } 00:09:31.183 ] 00:09:31.183 }' 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:31.183 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.440 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:31.440 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.440 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.440 [2024-11-20 15:57:29.642530] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:31.440 [2024-11-20 15:57:29.642702] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:31.440 [2024-11-20 15:57:29.645787] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:31.440 [2024-11-20 15:57:29.645917] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:31.441 [2024-11-20 15:57:29.646003] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:31.441 [2024-11-20 15:57:29.646075] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:09:31.441 { 00:09:31.441 "results": [ 00:09:31.441 { 00:09:31.441 "job": "raid_bdev1", 00:09:31.441 "core_mask": "0x1", 00:09:31.441 "workload": "randrw", 00:09:31.441 "percentage": 50, 00:09:31.441 "status": "finished", 00:09:31.441 "queue_depth": 1, 00:09:31.441 "io_size": 131072, 00:09:31.441 "runtime": 1.265214, 00:09:31.441 "iops": 20109.641530997917, 00:09:31.441 "mibps": 2513.7051913747396, 00:09:31.441 "io_failed": 0, 00:09:31.441 "io_timeout": 0, 00:09:31.441 "avg_latency_us": 46.53102373631557, 00:09:31.441 "min_latency_us": 28.75076923076923, 00:09:31.441 "max_latency_us": 1688.8123076923077 00:09:31.441 } 00:09:31.441 ], 00:09:31.441 "core_count": 1 00:09:31.441 } 00:09:31.441 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.441 15:57:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 62201 00:09:31.441 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 62201 ']' 00:09:31.441 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 62201 00:09:31.441 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:09:31.441 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:31.441 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 62201 00:09:31.441 killing process with pid 62201 00:09:31.441 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:31.441 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:31.441 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 62201' 00:09:31.441 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 62201 00:09:31.441 [2024-11-20 15:57:29.672304] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:31.441 15:57:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 62201 00:09:31.698 [2024-11-20 15:57:29.757640] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:32.266 15:57:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:32.266 15:57:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.XmEHGKIEV6 00:09:32.266 15:57:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:32.266 ************************************ 00:09:32.266 END TEST raid_write_error_test 00:09:32.266 ************************************ 00:09:32.266 15:57:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:32.266 15:57:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:32.266 15:57:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:32.266 15:57:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:32.266 15:57:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:32.266 00:09:32.266 real 0m3.416s 00:09:32.266 user 0m4.116s 00:09:32.266 sys 0m0.361s 00:09:32.266 15:57:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:32.266 15:57:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.266 15:57:30 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:09:32.266 15:57:30 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:32.266 15:57:30 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 3 false 00:09:32.266 15:57:30 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:32.266 15:57:30 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:32.266 15:57:30 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:32.266 ************************************ 00:09:32.266 START TEST raid_state_function_test 00:09:32.266 ************************************ 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 3 false 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:32.266 Process raid pid: 62328 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=62328 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 62328' 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 62328 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 62328 ']' 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:32.266 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.266 15:57:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:32.526 [2024-11-20 15:57:30.515244] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:09:32.526 [2024-11-20 15:57:30.515339] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:32.526 [2024-11-20 15:57:30.665447] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:32.526 [2024-11-20 15:57:30.748683] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:32.792 [2024-11-20 15:57:30.859485] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:32.792 [2024-11-20 15:57:30.859520] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.402 [2024-11-20 15:57:31.386433] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:33.402 [2024-11-20 15:57:31.386487] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:33.402 [2024-11-20 15:57:31.386495] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:33.402 [2024-11-20 15:57:31.386503] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:33.402 [2024-11-20 15:57:31.386508] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:33.402 [2024-11-20 15:57:31.386515] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:33.402 "name": "Existed_Raid", 00:09:33.402 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:33.402 "strip_size_kb": 64, 00:09:33.402 "state": "configuring", 00:09:33.402 "raid_level": "raid0", 00:09:33.402 "superblock": false, 00:09:33.402 "num_base_bdevs": 3, 00:09:33.402 "num_base_bdevs_discovered": 0, 00:09:33.402 "num_base_bdevs_operational": 3, 00:09:33.402 "base_bdevs_list": [ 00:09:33.402 { 00:09:33.402 "name": "BaseBdev1", 00:09:33.402 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:33.402 "is_configured": false, 00:09:33.402 "data_offset": 0, 00:09:33.402 "data_size": 0 00:09:33.402 }, 00:09:33.402 { 00:09:33.402 "name": "BaseBdev2", 00:09:33.402 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:33.402 "is_configured": false, 00:09:33.402 "data_offset": 0, 00:09:33.402 "data_size": 0 00:09:33.402 }, 00:09:33.402 { 00:09:33.402 "name": "BaseBdev3", 00:09:33.402 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:33.402 "is_configured": false, 00:09:33.402 "data_offset": 0, 00:09:33.402 "data_size": 0 00:09:33.402 } 00:09:33.402 ] 00:09:33.402 }' 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:33.402 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.660 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:33.660 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.660 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.660 [2024-11-20 15:57:31.694431] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:33.660 [2024-11-20 15:57:31.694462] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:09:33.660 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.660 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:33.660 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.660 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.660 [2024-11-20 15:57:31.702449] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:33.661 [2024-11-20 15:57:31.702490] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:33.661 [2024-11-20 15:57:31.702498] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:33.661 [2024-11-20 15:57:31.702504] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:33.661 [2024-11-20 15:57:31.702509] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:33.661 [2024-11-20 15:57:31.702516] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.661 [2024-11-20 15:57:31.730664] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:33.661 BaseBdev1 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.661 [ 00:09:33.661 { 00:09:33.661 "name": "BaseBdev1", 00:09:33.661 "aliases": [ 00:09:33.661 "20ecc70b-b8ab-4a00-9acb-742d0d9e5579" 00:09:33.661 ], 00:09:33.661 "product_name": "Malloc disk", 00:09:33.661 "block_size": 512, 00:09:33.661 "num_blocks": 65536, 00:09:33.661 "uuid": "20ecc70b-b8ab-4a00-9acb-742d0d9e5579", 00:09:33.661 "assigned_rate_limits": { 00:09:33.661 "rw_ios_per_sec": 0, 00:09:33.661 "rw_mbytes_per_sec": 0, 00:09:33.661 "r_mbytes_per_sec": 0, 00:09:33.661 "w_mbytes_per_sec": 0 00:09:33.661 }, 00:09:33.661 "claimed": true, 00:09:33.661 "claim_type": "exclusive_write", 00:09:33.661 "zoned": false, 00:09:33.661 "supported_io_types": { 00:09:33.661 "read": true, 00:09:33.661 "write": true, 00:09:33.661 "unmap": true, 00:09:33.661 "flush": true, 00:09:33.661 "reset": true, 00:09:33.661 "nvme_admin": false, 00:09:33.661 "nvme_io": false, 00:09:33.661 "nvme_io_md": false, 00:09:33.661 "write_zeroes": true, 00:09:33.661 "zcopy": true, 00:09:33.661 "get_zone_info": false, 00:09:33.661 "zone_management": false, 00:09:33.661 "zone_append": false, 00:09:33.661 "compare": false, 00:09:33.661 "compare_and_write": false, 00:09:33.661 "abort": true, 00:09:33.661 "seek_hole": false, 00:09:33.661 "seek_data": false, 00:09:33.661 "copy": true, 00:09:33.661 "nvme_iov_md": false 00:09:33.661 }, 00:09:33.661 "memory_domains": [ 00:09:33.661 { 00:09:33.661 "dma_device_id": "system", 00:09:33.661 "dma_device_type": 1 00:09:33.661 }, 00:09:33.661 { 00:09:33.661 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:33.661 "dma_device_type": 2 00:09:33.661 } 00:09:33.661 ], 00:09:33.661 "driver_specific": {} 00:09:33.661 } 00:09:33.661 ] 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:33.661 "name": "Existed_Raid", 00:09:33.661 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:33.661 "strip_size_kb": 64, 00:09:33.661 "state": "configuring", 00:09:33.661 "raid_level": "raid0", 00:09:33.661 "superblock": false, 00:09:33.661 "num_base_bdevs": 3, 00:09:33.661 "num_base_bdevs_discovered": 1, 00:09:33.661 "num_base_bdevs_operational": 3, 00:09:33.661 "base_bdevs_list": [ 00:09:33.661 { 00:09:33.661 "name": "BaseBdev1", 00:09:33.661 "uuid": "20ecc70b-b8ab-4a00-9acb-742d0d9e5579", 00:09:33.661 "is_configured": true, 00:09:33.661 "data_offset": 0, 00:09:33.661 "data_size": 65536 00:09:33.661 }, 00:09:33.661 { 00:09:33.661 "name": "BaseBdev2", 00:09:33.661 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:33.661 "is_configured": false, 00:09:33.661 "data_offset": 0, 00:09:33.661 "data_size": 0 00:09:33.661 }, 00:09:33.661 { 00:09:33.661 "name": "BaseBdev3", 00:09:33.661 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:33.661 "is_configured": false, 00:09:33.661 "data_offset": 0, 00:09:33.661 "data_size": 0 00:09:33.661 } 00:09:33.661 ] 00:09:33.661 }' 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:33.661 15:57:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.918 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:33.918 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.918 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.918 [2024-11-20 15:57:32.078779] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:33.919 [2024-11-20 15:57:32.078951] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.919 [2024-11-20 15:57:32.090860] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:33.919 [2024-11-20 15:57:32.092462] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:33.919 [2024-11-20 15:57:32.092505] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:33.919 [2024-11-20 15:57:32.092513] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:33.919 [2024-11-20 15:57:32.092520] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:33.919 "name": "Existed_Raid", 00:09:33.919 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:33.919 "strip_size_kb": 64, 00:09:33.919 "state": "configuring", 00:09:33.919 "raid_level": "raid0", 00:09:33.919 "superblock": false, 00:09:33.919 "num_base_bdevs": 3, 00:09:33.919 "num_base_bdevs_discovered": 1, 00:09:33.919 "num_base_bdevs_operational": 3, 00:09:33.919 "base_bdevs_list": [ 00:09:33.919 { 00:09:33.919 "name": "BaseBdev1", 00:09:33.919 "uuid": "20ecc70b-b8ab-4a00-9acb-742d0d9e5579", 00:09:33.919 "is_configured": true, 00:09:33.919 "data_offset": 0, 00:09:33.919 "data_size": 65536 00:09:33.919 }, 00:09:33.919 { 00:09:33.919 "name": "BaseBdev2", 00:09:33.919 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:33.919 "is_configured": false, 00:09:33.919 "data_offset": 0, 00:09:33.919 "data_size": 0 00:09:33.919 }, 00:09:33.919 { 00:09:33.919 "name": "BaseBdev3", 00:09:33.919 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:33.919 "is_configured": false, 00:09:33.919 "data_offset": 0, 00:09:33.919 "data_size": 0 00:09:33.919 } 00:09:33.919 ] 00:09:33.919 }' 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:33.919 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.177 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:34.177 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.177 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.434 [2024-11-20 15:57:32.441480] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:34.434 BaseBdev2 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.434 [ 00:09:34.434 { 00:09:34.434 "name": "BaseBdev2", 00:09:34.434 "aliases": [ 00:09:34.434 "aba73bee-3473-43b5-8b7f-a7f68beb581d" 00:09:34.434 ], 00:09:34.434 "product_name": "Malloc disk", 00:09:34.434 "block_size": 512, 00:09:34.434 "num_blocks": 65536, 00:09:34.434 "uuid": "aba73bee-3473-43b5-8b7f-a7f68beb581d", 00:09:34.434 "assigned_rate_limits": { 00:09:34.434 "rw_ios_per_sec": 0, 00:09:34.434 "rw_mbytes_per_sec": 0, 00:09:34.434 "r_mbytes_per_sec": 0, 00:09:34.434 "w_mbytes_per_sec": 0 00:09:34.434 }, 00:09:34.434 "claimed": true, 00:09:34.434 "claim_type": "exclusive_write", 00:09:34.434 "zoned": false, 00:09:34.434 "supported_io_types": { 00:09:34.434 "read": true, 00:09:34.434 "write": true, 00:09:34.434 "unmap": true, 00:09:34.434 "flush": true, 00:09:34.434 "reset": true, 00:09:34.434 "nvme_admin": false, 00:09:34.434 "nvme_io": false, 00:09:34.434 "nvme_io_md": false, 00:09:34.434 "write_zeroes": true, 00:09:34.434 "zcopy": true, 00:09:34.434 "get_zone_info": false, 00:09:34.434 "zone_management": false, 00:09:34.434 "zone_append": false, 00:09:34.434 "compare": false, 00:09:34.434 "compare_and_write": false, 00:09:34.434 "abort": true, 00:09:34.434 "seek_hole": false, 00:09:34.434 "seek_data": false, 00:09:34.434 "copy": true, 00:09:34.434 "nvme_iov_md": false 00:09:34.434 }, 00:09:34.434 "memory_domains": [ 00:09:34.434 { 00:09:34.434 "dma_device_id": "system", 00:09:34.434 "dma_device_type": 1 00:09:34.434 }, 00:09:34.434 { 00:09:34.434 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.434 "dma_device_type": 2 00:09:34.434 } 00:09:34.434 ], 00:09:34.434 "driver_specific": {} 00:09:34.434 } 00:09:34.434 ] 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:34.434 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:34.435 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:34.435 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:34.435 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:34.435 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:34.435 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.435 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.435 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.435 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:34.435 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.435 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:34.435 "name": "Existed_Raid", 00:09:34.435 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:34.435 "strip_size_kb": 64, 00:09:34.435 "state": "configuring", 00:09:34.435 "raid_level": "raid0", 00:09:34.435 "superblock": false, 00:09:34.435 "num_base_bdevs": 3, 00:09:34.435 "num_base_bdevs_discovered": 2, 00:09:34.435 "num_base_bdevs_operational": 3, 00:09:34.435 "base_bdevs_list": [ 00:09:34.435 { 00:09:34.435 "name": "BaseBdev1", 00:09:34.435 "uuid": "20ecc70b-b8ab-4a00-9acb-742d0d9e5579", 00:09:34.435 "is_configured": true, 00:09:34.435 "data_offset": 0, 00:09:34.435 "data_size": 65536 00:09:34.435 }, 00:09:34.435 { 00:09:34.435 "name": "BaseBdev2", 00:09:34.435 "uuid": "aba73bee-3473-43b5-8b7f-a7f68beb581d", 00:09:34.435 "is_configured": true, 00:09:34.435 "data_offset": 0, 00:09:34.435 "data_size": 65536 00:09:34.435 }, 00:09:34.435 { 00:09:34.435 "name": "BaseBdev3", 00:09:34.435 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:34.435 "is_configured": false, 00:09:34.435 "data_offset": 0, 00:09:34.435 "data_size": 0 00:09:34.435 } 00:09:34.435 ] 00:09:34.435 }' 00:09:34.435 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:34.435 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.691 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:34.691 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.691 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.691 [2024-11-20 15:57:32.839195] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:34.691 [2024-11-20 15:57:32.839388] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:34.691 [2024-11-20 15:57:32.839421] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:09:34.691 [2024-11-20 15:57:32.839703] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:09:34.691 [2024-11-20 15:57:32.839892] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:34.692 [2024-11-20 15:57:32.839953] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:09:34.692 [2024-11-20 15:57:32.840213] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:34.692 BaseBdev3 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.692 [ 00:09:34.692 { 00:09:34.692 "name": "BaseBdev3", 00:09:34.692 "aliases": [ 00:09:34.692 "bfc1adb1-89c0-4cbc-bca4-fa85afe529cc" 00:09:34.692 ], 00:09:34.692 "product_name": "Malloc disk", 00:09:34.692 "block_size": 512, 00:09:34.692 "num_blocks": 65536, 00:09:34.692 "uuid": "bfc1adb1-89c0-4cbc-bca4-fa85afe529cc", 00:09:34.692 "assigned_rate_limits": { 00:09:34.692 "rw_ios_per_sec": 0, 00:09:34.692 "rw_mbytes_per_sec": 0, 00:09:34.692 "r_mbytes_per_sec": 0, 00:09:34.692 "w_mbytes_per_sec": 0 00:09:34.692 }, 00:09:34.692 "claimed": true, 00:09:34.692 "claim_type": "exclusive_write", 00:09:34.692 "zoned": false, 00:09:34.692 "supported_io_types": { 00:09:34.692 "read": true, 00:09:34.692 "write": true, 00:09:34.692 "unmap": true, 00:09:34.692 "flush": true, 00:09:34.692 "reset": true, 00:09:34.692 "nvme_admin": false, 00:09:34.692 "nvme_io": false, 00:09:34.692 "nvme_io_md": false, 00:09:34.692 "write_zeroes": true, 00:09:34.692 "zcopy": true, 00:09:34.692 "get_zone_info": false, 00:09:34.692 "zone_management": false, 00:09:34.692 "zone_append": false, 00:09:34.692 "compare": false, 00:09:34.692 "compare_and_write": false, 00:09:34.692 "abort": true, 00:09:34.692 "seek_hole": false, 00:09:34.692 "seek_data": false, 00:09:34.692 "copy": true, 00:09:34.692 "nvme_iov_md": false 00:09:34.692 }, 00:09:34.692 "memory_domains": [ 00:09:34.692 { 00:09:34.692 "dma_device_id": "system", 00:09:34.692 "dma_device_type": 1 00:09:34.692 }, 00:09:34.692 { 00:09:34.692 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.692 "dma_device_type": 2 00:09:34.692 } 00:09:34.692 ], 00:09:34.692 "driver_specific": {} 00:09:34.692 } 00:09:34.692 ] 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:34.692 "name": "Existed_Raid", 00:09:34.692 "uuid": "e22dd55f-3868-4d69-b9dd-d26696ae3a7c", 00:09:34.692 "strip_size_kb": 64, 00:09:34.692 "state": "online", 00:09:34.692 "raid_level": "raid0", 00:09:34.692 "superblock": false, 00:09:34.692 "num_base_bdevs": 3, 00:09:34.692 "num_base_bdevs_discovered": 3, 00:09:34.692 "num_base_bdevs_operational": 3, 00:09:34.692 "base_bdevs_list": [ 00:09:34.692 { 00:09:34.692 "name": "BaseBdev1", 00:09:34.692 "uuid": "20ecc70b-b8ab-4a00-9acb-742d0d9e5579", 00:09:34.692 "is_configured": true, 00:09:34.692 "data_offset": 0, 00:09:34.692 "data_size": 65536 00:09:34.692 }, 00:09:34.692 { 00:09:34.692 "name": "BaseBdev2", 00:09:34.692 "uuid": "aba73bee-3473-43b5-8b7f-a7f68beb581d", 00:09:34.692 "is_configured": true, 00:09:34.692 "data_offset": 0, 00:09:34.692 "data_size": 65536 00:09:34.692 }, 00:09:34.692 { 00:09:34.692 "name": "BaseBdev3", 00:09:34.692 "uuid": "bfc1adb1-89c0-4cbc-bca4-fa85afe529cc", 00:09:34.692 "is_configured": true, 00:09:34.692 "data_offset": 0, 00:09:34.692 "data_size": 65536 00:09:34.692 } 00:09:34.692 ] 00:09:34.692 }' 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:34.692 15:57:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.949 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:34.949 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:34.949 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:34.949 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:34.949 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:34.949 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:34.949 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:34.949 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:34.949 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.949 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.949 [2024-11-20 15:57:33.195572] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:35.207 "name": "Existed_Raid", 00:09:35.207 "aliases": [ 00:09:35.207 "e22dd55f-3868-4d69-b9dd-d26696ae3a7c" 00:09:35.207 ], 00:09:35.207 "product_name": "Raid Volume", 00:09:35.207 "block_size": 512, 00:09:35.207 "num_blocks": 196608, 00:09:35.207 "uuid": "e22dd55f-3868-4d69-b9dd-d26696ae3a7c", 00:09:35.207 "assigned_rate_limits": { 00:09:35.207 "rw_ios_per_sec": 0, 00:09:35.207 "rw_mbytes_per_sec": 0, 00:09:35.207 "r_mbytes_per_sec": 0, 00:09:35.207 "w_mbytes_per_sec": 0 00:09:35.207 }, 00:09:35.207 "claimed": false, 00:09:35.207 "zoned": false, 00:09:35.207 "supported_io_types": { 00:09:35.207 "read": true, 00:09:35.207 "write": true, 00:09:35.207 "unmap": true, 00:09:35.207 "flush": true, 00:09:35.207 "reset": true, 00:09:35.207 "nvme_admin": false, 00:09:35.207 "nvme_io": false, 00:09:35.207 "nvme_io_md": false, 00:09:35.207 "write_zeroes": true, 00:09:35.207 "zcopy": false, 00:09:35.207 "get_zone_info": false, 00:09:35.207 "zone_management": false, 00:09:35.207 "zone_append": false, 00:09:35.207 "compare": false, 00:09:35.207 "compare_and_write": false, 00:09:35.207 "abort": false, 00:09:35.207 "seek_hole": false, 00:09:35.207 "seek_data": false, 00:09:35.207 "copy": false, 00:09:35.207 "nvme_iov_md": false 00:09:35.207 }, 00:09:35.207 "memory_domains": [ 00:09:35.207 { 00:09:35.207 "dma_device_id": "system", 00:09:35.207 "dma_device_type": 1 00:09:35.207 }, 00:09:35.207 { 00:09:35.207 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:35.207 "dma_device_type": 2 00:09:35.207 }, 00:09:35.207 { 00:09:35.207 "dma_device_id": "system", 00:09:35.207 "dma_device_type": 1 00:09:35.207 }, 00:09:35.207 { 00:09:35.207 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:35.207 "dma_device_type": 2 00:09:35.207 }, 00:09:35.207 { 00:09:35.207 "dma_device_id": "system", 00:09:35.207 "dma_device_type": 1 00:09:35.207 }, 00:09:35.207 { 00:09:35.207 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:35.207 "dma_device_type": 2 00:09:35.207 } 00:09:35.207 ], 00:09:35.207 "driver_specific": { 00:09:35.207 "raid": { 00:09:35.207 "uuid": "e22dd55f-3868-4d69-b9dd-d26696ae3a7c", 00:09:35.207 "strip_size_kb": 64, 00:09:35.207 "state": "online", 00:09:35.207 "raid_level": "raid0", 00:09:35.207 "superblock": false, 00:09:35.207 "num_base_bdevs": 3, 00:09:35.207 "num_base_bdevs_discovered": 3, 00:09:35.207 "num_base_bdevs_operational": 3, 00:09:35.207 "base_bdevs_list": [ 00:09:35.207 { 00:09:35.207 "name": "BaseBdev1", 00:09:35.207 "uuid": "20ecc70b-b8ab-4a00-9acb-742d0d9e5579", 00:09:35.207 "is_configured": true, 00:09:35.207 "data_offset": 0, 00:09:35.207 "data_size": 65536 00:09:35.207 }, 00:09:35.207 { 00:09:35.207 "name": "BaseBdev2", 00:09:35.207 "uuid": "aba73bee-3473-43b5-8b7f-a7f68beb581d", 00:09:35.207 "is_configured": true, 00:09:35.207 "data_offset": 0, 00:09:35.207 "data_size": 65536 00:09:35.207 }, 00:09:35.207 { 00:09:35.207 "name": "BaseBdev3", 00:09:35.207 "uuid": "bfc1adb1-89c0-4cbc-bca4-fa85afe529cc", 00:09:35.207 "is_configured": true, 00:09:35.207 "data_offset": 0, 00:09:35.207 "data_size": 65536 00:09:35.207 } 00:09:35.207 ] 00:09:35.207 } 00:09:35.207 } 00:09:35.207 }' 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:35.207 BaseBdev2 00:09:35.207 BaseBdev3' 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.207 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.208 [2024-11-20 15:57:33.367376] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:35.208 [2024-11-20 15:57:33.367399] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:35.208 [2024-11-20 15:57:33.367439] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:35.208 "name": "Existed_Raid", 00:09:35.208 "uuid": "e22dd55f-3868-4d69-b9dd-d26696ae3a7c", 00:09:35.208 "strip_size_kb": 64, 00:09:35.208 "state": "offline", 00:09:35.208 "raid_level": "raid0", 00:09:35.208 "superblock": false, 00:09:35.208 "num_base_bdevs": 3, 00:09:35.208 "num_base_bdevs_discovered": 2, 00:09:35.208 "num_base_bdevs_operational": 2, 00:09:35.208 "base_bdevs_list": [ 00:09:35.208 { 00:09:35.208 "name": null, 00:09:35.208 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:35.208 "is_configured": false, 00:09:35.208 "data_offset": 0, 00:09:35.208 "data_size": 65536 00:09:35.208 }, 00:09:35.208 { 00:09:35.208 "name": "BaseBdev2", 00:09:35.208 "uuid": "aba73bee-3473-43b5-8b7f-a7f68beb581d", 00:09:35.208 "is_configured": true, 00:09:35.208 "data_offset": 0, 00:09:35.208 "data_size": 65536 00:09:35.208 }, 00:09:35.208 { 00:09:35.208 "name": "BaseBdev3", 00:09:35.208 "uuid": "bfc1adb1-89c0-4cbc-bca4-fa85afe529cc", 00:09:35.208 "is_configured": true, 00:09:35.208 "data_offset": 0, 00:09:35.208 "data_size": 65536 00:09:35.208 } 00:09:35.208 ] 00:09:35.208 }' 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:35.208 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.466 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:35.466 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:35.466 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:35.466 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.466 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.466 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.723 [2024-11-20 15:57:33.742947] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.723 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.724 [2024-11-20 15:57:33.829423] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:35.724 [2024-11-20 15:57:33.829548] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.724 BaseBdev2 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.724 [ 00:09:35.724 { 00:09:35.724 "name": "BaseBdev2", 00:09:35.724 "aliases": [ 00:09:35.724 "038100a5-75b2-4d72-b1fc-7156862bb82e" 00:09:35.724 ], 00:09:35.724 "product_name": "Malloc disk", 00:09:35.724 "block_size": 512, 00:09:35.724 "num_blocks": 65536, 00:09:35.724 "uuid": "038100a5-75b2-4d72-b1fc-7156862bb82e", 00:09:35.724 "assigned_rate_limits": { 00:09:35.724 "rw_ios_per_sec": 0, 00:09:35.724 "rw_mbytes_per_sec": 0, 00:09:35.724 "r_mbytes_per_sec": 0, 00:09:35.724 "w_mbytes_per_sec": 0 00:09:35.724 }, 00:09:35.724 "claimed": false, 00:09:35.724 "zoned": false, 00:09:35.724 "supported_io_types": { 00:09:35.724 "read": true, 00:09:35.724 "write": true, 00:09:35.724 "unmap": true, 00:09:35.724 "flush": true, 00:09:35.724 "reset": true, 00:09:35.724 "nvme_admin": false, 00:09:35.724 "nvme_io": false, 00:09:35.724 "nvme_io_md": false, 00:09:35.724 "write_zeroes": true, 00:09:35.724 "zcopy": true, 00:09:35.724 "get_zone_info": false, 00:09:35.724 "zone_management": false, 00:09:35.724 "zone_append": false, 00:09:35.724 "compare": false, 00:09:35.724 "compare_and_write": false, 00:09:35.724 "abort": true, 00:09:35.724 "seek_hole": false, 00:09:35.724 "seek_data": false, 00:09:35.724 "copy": true, 00:09:35.724 "nvme_iov_md": false 00:09:35.724 }, 00:09:35.724 "memory_domains": [ 00:09:35.724 { 00:09:35.724 "dma_device_id": "system", 00:09:35.724 "dma_device_type": 1 00:09:35.724 }, 00:09:35.724 { 00:09:35.724 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:35.724 "dma_device_type": 2 00:09:35.724 } 00:09:35.724 ], 00:09:35.724 "driver_specific": {} 00:09:35.724 } 00:09:35.724 ] 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.724 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.983 BaseBdev3 00:09:35.983 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.983 15:57:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:35.983 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:35.983 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:35.983 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:35.983 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:35.983 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:35.983 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:35.983 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.983 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.983 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.983 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:35.983 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.983 15:57:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.983 [ 00:09:35.983 { 00:09:35.983 "name": "BaseBdev3", 00:09:35.983 "aliases": [ 00:09:35.983 "1488d52b-3f6a-49f0-b7c7-1ce28accea30" 00:09:35.983 ], 00:09:35.983 "product_name": "Malloc disk", 00:09:35.983 "block_size": 512, 00:09:35.983 "num_blocks": 65536, 00:09:35.983 "uuid": "1488d52b-3f6a-49f0-b7c7-1ce28accea30", 00:09:35.983 "assigned_rate_limits": { 00:09:35.983 "rw_ios_per_sec": 0, 00:09:35.983 "rw_mbytes_per_sec": 0, 00:09:35.983 "r_mbytes_per_sec": 0, 00:09:35.983 "w_mbytes_per_sec": 0 00:09:35.983 }, 00:09:35.983 "claimed": false, 00:09:35.983 "zoned": false, 00:09:35.983 "supported_io_types": { 00:09:35.983 "read": true, 00:09:35.983 "write": true, 00:09:35.983 "unmap": true, 00:09:35.983 "flush": true, 00:09:35.983 "reset": true, 00:09:35.983 "nvme_admin": false, 00:09:35.983 "nvme_io": false, 00:09:35.983 "nvme_io_md": false, 00:09:35.983 "write_zeroes": true, 00:09:35.983 "zcopy": true, 00:09:35.983 "get_zone_info": false, 00:09:35.983 "zone_management": false, 00:09:35.983 "zone_append": false, 00:09:35.983 "compare": false, 00:09:35.983 "compare_and_write": false, 00:09:35.983 "abort": true, 00:09:35.983 "seek_hole": false, 00:09:35.983 "seek_data": false, 00:09:35.983 "copy": true, 00:09:35.983 "nvme_iov_md": false 00:09:35.983 }, 00:09:35.983 "memory_domains": [ 00:09:35.983 { 00:09:35.983 "dma_device_id": "system", 00:09:35.983 "dma_device_type": 1 00:09:35.983 }, 00:09:35.983 { 00:09:35.983 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:35.983 "dma_device_type": 2 00:09:35.983 } 00:09:35.983 ], 00:09:35.983 "driver_specific": {} 00:09:35.983 } 00:09:35.983 ] 00:09:35.983 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.983 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:35.983 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:35.983 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:35.983 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:35.983 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.984 [2024-11-20 15:57:34.011791] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:35.984 [2024-11-20 15:57:34.011911] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:35.984 [2024-11-20 15:57:34.011937] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:35.984 [2024-11-20 15:57:34.013481] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:35.984 "name": "Existed_Raid", 00:09:35.984 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:35.984 "strip_size_kb": 64, 00:09:35.984 "state": "configuring", 00:09:35.984 "raid_level": "raid0", 00:09:35.984 "superblock": false, 00:09:35.984 "num_base_bdevs": 3, 00:09:35.984 "num_base_bdevs_discovered": 2, 00:09:35.984 "num_base_bdevs_operational": 3, 00:09:35.984 "base_bdevs_list": [ 00:09:35.984 { 00:09:35.984 "name": "BaseBdev1", 00:09:35.984 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:35.984 "is_configured": false, 00:09:35.984 "data_offset": 0, 00:09:35.984 "data_size": 0 00:09:35.984 }, 00:09:35.984 { 00:09:35.984 "name": "BaseBdev2", 00:09:35.984 "uuid": "038100a5-75b2-4d72-b1fc-7156862bb82e", 00:09:35.984 "is_configured": true, 00:09:35.984 "data_offset": 0, 00:09:35.984 "data_size": 65536 00:09:35.984 }, 00:09:35.984 { 00:09:35.984 "name": "BaseBdev3", 00:09:35.984 "uuid": "1488d52b-3f6a-49f0-b7c7-1ce28accea30", 00:09:35.984 "is_configured": true, 00:09:35.984 "data_offset": 0, 00:09:35.984 "data_size": 65536 00:09:35.984 } 00:09:35.984 ] 00:09:35.984 }' 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:35.984 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.307 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:36.307 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.307 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.307 [2024-11-20 15:57:34.323866] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:36.307 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.307 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:36.307 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:36.307 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:36.308 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:36.308 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:36.308 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:36.308 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.308 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.308 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.308 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.308 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:36.308 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.308 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.308 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.308 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.308 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:36.308 "name": "Existed_Raid", 00:09:36.308 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:36.308 "strip_size_kb": 64, 00:09:36.308 "state": "configuring", 00:09:36.308 "raid_level": "raid0", 00:09:36.308 "superblock": false, 00:09:36.308 "num_base_bdevs": 3, 00:09:36.308 "num_base_bdevs_discovered": 1, 00:09:36.308 "num_base_bdevs_operational": 3, 00:09:36.308 "base_bdevs_list": [ 00:09:36.308 { 00:09:36.308 "name": "BaseBdev1", 00:09:36.308 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:36.308 "is_configured": false, 00:09:36.308 "data_offset": 0, 00:09:36.308 "data_size": 0 00:09:36.308 }, 00:09:36.308 { 00:09:36.308 "name": null, 00:09:36.308 "uuid": "038100a5-75b2-4d72-b1fc-7156862bb82e", 00:09:36.308 "is_configured": false, 00:09:36.308 "data_offset": 0, 00:09:36.308 "data_size": 65536 00:09:36.308 }, 00:09:36.308 { 00:09:36.308 "name": "BaseBdev3", 00:09:36.308 "uuid": "1488d52b-3f6a-49f0-b7c7-1ce28accea30", 00:09:36.308 "is_configured": true, 00:09:36.308 "data_offset": 0, 00:09:36.308 "data_size": 65536 00:09:36.308 } 00:09:36.308 ] 00:09:36.308 }' 00:09:36.308 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:36.308 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.581 [2024-11-20 15:57:34.694237] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:36.581 BaseBdev1 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.581 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.581 [ 00:09:36.581 { 00:09:36.581 "name": "BaseBdev1", 00:09:36.581 "aliases": [ 00:09:36.581 "a7f3090e-09d7-494b-b316-3382977600e8" 00:09:36.581 ], 00:09:36.581 "product_name": "Malloc disk", 00:09:36.581 "block_size": 512, 00:09:36.581 "num_blocks": 65536, 00:09:36.581 "uuid": "a7f3090e-09d7-494b-b316-3382977600e8", 00:09:36.581 "assigned_rate_limits": { 00:09:36.581 "rw_ios_per_sec": 0, 00:09:36.581 "rw_mbytes_per_sec": 0, 00:09:36.581 "r_mbytes_per_sec": 0, 00:09:36.581 "w_mbytes_per_sec": 0 00:09:36.581 }, 00:09:36.581 "claimed": true, 00:09:36.581 "claim_type": "exclusive_write", 00:09:36.581 "zoned": false, 00:09:36.581 "supported_io_types": { 00:09:36.581 "read": true, 00:09:36.581 "write": true, 00:09:36.581 "unmap": true, 00:09:36.581 "flush": true, 00:09:36.581 "reset": true, 00:09:36.581 "nvme_admin": false, 00:09:36.581 "nvme_io": false, 00:09:36.581 "nvme_io_md": false, 00:09:36.581 "write_zeroes": true, 00:09:36.581 "zcopy": true, 00:09:36.581 "get_zone_info": false, 00:09:36.581 "zone_management": false, 00:09:36.581 "zone_append": false, 00:09:36.581 "compare": false, 00:09:36.581 "compare_and_write": false, 00:09:36.581 "abort": true, 00:09:36.581 "seek_hole": false, 00:09:36.581 "seek_data": false, 00:09:36.581 "copy": true, 00:09:36.581 "nvme_iov_md": false 00:09:36.581 }, 00:09:36.582 "memory_domains": [ 00:09:36.582 { 00:09:36.582 "dma_device_id": "system", 00:09:36.582 "dma_device_type": 1 00:09:36.582 }, 00:09:36.582 { 00:09:36.582 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:36.582 "dma_device_type": 2 00:09:36.582 } 00:09:36.582 ], 00:09:36.582 "driver_specific": {} 00:09:36.582 } 00:09:36.582 ] 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:36.582 "name": "Existed_Raid", 00:09:36.582 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:36.582 "strip_size_kb": 64, 00:09:36.582 "state": "configuring", 00:09:36.582 "raid_level": "raid0", 00:09:36.582 "superblock": false, 00:09:36.582 "num_base_bdevs": 3, 00:09:36.582 "num_base_bdevs_discovered": 2, 00:09:36.582 "num_base_bdevs_operational": 3, 00:09:36.582 "base_bdevs_list": [ 00:09:36.582 { 00:09:36.582 "name": "BaseBdev1", 00:09:36.582 "uuid": "a7f3090e-09d7-494b-b316-3382977600e8", 00:09:36.582 "is_configured": true, 00:09:36.582 "data_offset": 0, 00:09:36.582 "data_size": 65536 00:09:36.582 }, 00:09:36.582 { 00:09:36.582 "name": null, 00:09:36.582 "uuid": "038100a5-75b2-4d72-b1fc-7156862bb82e", 00:09:36.582 "is_configured": false, 00:09:36.582 "data_offset": 0, 00:09:36.582 "data_size": 65536 00:09:36.582 }, 00:09:36.582 { 00:09:36.582 "name": "BaseBdev3", 00:09:36.582 "uuid": "1488d52b-3f6a-49f0-b7c7-1ce28accea30", 00:09:36.582 "is_configured": true, 00:09:36.582 "data_offset": 0, 00:09:36.582 "data_size": 65536 00:09:36.582 } 00:09:36.582 ] 00:09:36.582 }' 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:36.582 15:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.839 [2024-11-20 15:57:35.054363] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.839 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.097 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:37.097 "name": "Existed_Raid", 00:09:37.097 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:37.097 "strip_size_kb": 64, 00:09:37.097 "state": "configuring", 00:09:37.097 "raid_level": "raid0", 00:09:37.097 "superblock": false, 00:09:37.097 "num_base_bdevs": 3, 00:09:37.097 "num_base_bdevs_discovered": 1, 00:09:37.097 "num_base_bdevs_operational": 3, 00:09:37.097 "base_bdevs_list": [ 00:09:37.097 { 00:09:37.097 "name": "BaseBdev1", 00:09:37.097 "uuid": "a7f3090e-09d7-494b-b316-3382977600e8", 00:09:37.097 "is_configured": true, 00:09:37.097 "data_offset": 0, 00:09:37.097 "data_size": 65536 00:09:37.097 }, 00:09:37.097 { 00:09:37.097 "name": null, 00:09:37.097 "uuid": "038100a5-75b2-4d72-b1fc-7156862bb82e", 00:09:37.097 "is_configured": false, 00:09:37.097 "data_offset": 0, 00:09:37.097 "data_size": 65536 00:09:37.097 }, 00:09:37.097 { 00:09:37.097 "name": null, 00:09:37.097 "uuid": "1488d52b-3f6a-49f0-b7c7-1ce28accea30", 00:09:37.097 "is_configured": false, 00:09:37.097 "data_offset": 0, 00:09:37.097 "data_size": 65536 00:09:37.097 } 00:09:37.097 ] 00:09:37.097 }' 00:09:37.097 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:37.097 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.355 [2024-11-20 15:57:35.398453] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:37.355 "name": "Existed_Raid", 00:09:37.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:37.355 "strip_size_kb": 64, 00:09:37.355 "state": "configuring", 00:09:37.355 "raid_level": "raid0", 00:09:37.355 "superblock": false, 00:09:37.355 "num_base_bdevs": 3, 00:09:37.355 "num_base_bdevs_discovered": 2, 00:09:37.355 "num_base_bdevs_operational": 3, 00:09:37.355 "base_bdevs_list": [ 00:09:37.355 { 00:09:37.355 "name": "BaseBdev1", 00:09:37.355 "uuid": "a7f3090e-09d7-494b-b316-3382977600e8", 00:09:37.355 "is_configured": true, 00:09:37.355 "data_offset": 0, 00:09:37.355 "data_size": 65536 00:09:37.355 }, 00:09:37.355 { 00:09:37.355 "name": null, 00:09:37.355 "uuid": "038100a5-75b2-4d72-b1fc-7156862bb82e", 00:09:37.355 "is_configured": false, 00:09:37.355 "data_offset": 0, 00:09:37.355 "data_size": 65536 00:09:37.355 }, 00:09:37.355 { 00:09:37.355 "name": "BaseBdev3", 00:09:37.355 "uuid": "1488d52b-3f6a-49f0-b7c7-1ce28accea30", 00:09:37.355 "is_configured": true, 00:09:37.355 "data_offset": 0, 00:09:37.355 "data_size": 65536 00:09:37.355 } 00:09:37.355 ] 00:09:37.355 }' 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:37.355 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.613 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:37.613 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.613 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.613 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.613 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.613 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:37.613 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:37.613 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.613 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.613 [2024-11-20 15:57:35.734520] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:37.613 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:37.614 "name": "Existed_Raid", 00:09:37.614 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:37.614 "strip_size_kb": 64, 00:09:37.614 "state": "configuring", 00:09:37.614 "raid_level": "raid0", 00:09:37.614 "superblock": false, 00:09:37.614 "num_base_bdevs": 3, 00:09:37.614 "num_base_bdevs_discovered": 1, 00:09:37.614 "num_base_bdevs_operational": 3, 00:09:37.614 "base_bdevs_list": [ 00:09:37.614 { 00:09:37.614 "name": null, 00:09:37.614 "uuid": "a7f3090e-09d7-494b-b316-3382977600e8", 00:09:37.614 "is_configured": false, 00:09:37.614 "data_offset": 0, 00:09:37.614 "data_size": 65536 00:09:37.614 }, 00:09:37.614 { 00:09:37.614 "name": null, 00:09:37.614 "uuid": "038100a5-75b2-4d72-b1fc-7156862bb82e", 00:09:37.614 "is_configured": false, 00:09:37.614 "data_offset": 0, 00:09:37.614 "data_size": 65536 00:09:37.614 }, 00:09:37.614 { 00:09:37.614 "name": "BaseBdev3", 00:09:37.614 "uuid": "1488d52b-3f6a-49f0-b7c7-1ce28accea30", 00:09:37.614 "is_configured": true, 00:09:37.614 "data_offset": 0, 00:09:37.614 "data_size": 65536 00:09:37.614 } 00:09:37.614 ] 00:09:37.614 }' 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:37.614 15:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.872 [2024-11-20 15:57:36.093610] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.872 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.130 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:38.130 "name": "Existed_Raid", 00:09:38.130 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:38.130 "strip_size_kb": 64, 00:09:38.130 "state": "configuring", 00:09:38.130 "raid_level": "raid0", 00:09:38.130 "superblock": false, 00:09:38.130 "num_base_bdevs": 3, 00:09:38.130 "num_base_bdevs_discovered": 2, 00:09:38.130 "num_base_bdevs_operational": 3, 00:09:38.130 "base_bdevs_list": [ 00:09:38.130 { 00:09:38.130 "name": null, 00:09:38.130 "uuid": "a7f3090e-09d7-494b-b316-3382977600e8", 00:09:38.130 "is_configured": false, 00:09:38.130 "data_offset": 0, 00:09:38.130 "data_size": 65536 00:09:38.130 }, 00:09:38.130 { 00:09:38.130 "name": "BaseBdev2", 00:09:38.130 "uuid": "038100a5-75b2-4d72-b1fc-7156862bb82e", 00:09:38.130 "is_configured": true, 00:09:38.130 "data_offset": 0, 00:09:38.130 "data_size": 65536 00:09:38.130 }, 00:09:38.130 { 00:09:38.130 "name": "BaseBdev3", 00:09:38.130 "uuid": "1488d52b-3f6a-49f0-b7c7-1ce28accea30", 00:09:38.130 "is_configured": true, 00:09:38.130 "data_offset": 0, 00:09:38.130 "data_size": 65536 00:09:38.130 } 00:09:38.130 ] 00:09:38.130 }' 00:09:38.130 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:38.130 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u a7f3090e-09d7-494b-b316-3382977600e8 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.389 [2024-11-20 15:57:36.463947] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:38.389 [2024-11-20 15:57:36.463978] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:09:38.389 [2024-11-20 15:57:36.463985] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:09:38.389 [2024-11-20 15:57:36.464173] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:09:38.389 [2024-11-20 15:57:36.464278] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:09:38.389 [2024-11-20 15:57:36.464284] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:09:38.389 [2024-11-20 15:57:36.464464] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:38.389 NewBaseBdev 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.389 [ 00:09:38.389 { 00:09:38.389 "name": "NewBaseBdev", 00:09:38.389 "aliases": [ 00:09:38.389 "a7f3090e-09d7-494b-b316-3382977600e8" 00:09:38.389 ], 00:09:38.389 "product_name": "Malloc disk", 00:09:38.389 "block_size": 512, 00:09:38.389 "num_blocks": 65536, 00:09:38.389 "uuid": "a7f3090e-09d7-494b-b316-3382977600e8", 00:09:38.389 "assigned_rate_limits": { 00:09:38.389 "rw_ios_per_sec": 0, 00:09:38.389 "rw_mbytes_per_sec": 0, 00:09:38.389 "r_mbytes_per_sec": 0, 00:09:38.389 "w_mbytes_per_sec": 0 00:09:38.389 }, 00:09:38.389 "claimed": true, 00:09:38.389 "claim_type": "exclusive_write", 00:09:38.389 "zoned": false, 00:09:38.389 "supported_io_types": { 00:09:38.389 "read": true, 00:09:38.389 "write": true, 00:09:38.389 "unmap": true, 00:09:38.389 "flush": true, 00:09:38.389 "reset": true, 00:09:38.389 "nvme_admin": false, 00:09:38.389 "nvme_io": false, 00:09:38.389 "nvme_io_md": false, 00:09:38.389 "write_zeroes": true, 00:09:38.389 "zcopy": true, 00:09:38.389 "get_zone_info": false, 00:09:38.389 "zone_management": false, 00:09:38.389 "zone_append": false, 00:09:38.389 "compare": false, 00:09:38.389 "compare_and_write": false, 00:09:38.389 "abort": true, 00:09:38.389 "seek_hole": false, 00:09:38.389 "seek_data": false, 00:09:38.389 "copy": true, 00:09:38.389 "nvme_iov_md": false 00:09:38.389 }, 00:09:38.389 "memory_domains": [ 00:09:38.389 { 00:09:38.389 "dma_device_id": "system", 00:09:38.389 "dma_device_type": 1 00:09:38.389 }, 00:09:38.389 { 00:09:38.389 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:38.389 "dma_device_type": 2 00:09:38.389 } 00:09:38.389 ], 00:09:38.389 "driver_specific": {} 00:09:38.389 } 00:09:38.389 ] 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:38.389 "name": "Existed_Raid", 00:09:38.389 "uuid": "6bf29723-dfa2-49e1-a182-1f85b4070e1d", 00:09:38.389 "strip_size_kb": 64, 00:09:38.389 "state": "online", 00:09:38.389 "raid_level": "raid0", 00:09:38.389 "superblock": false, 00:09:38.389 "num_base_bdevs": 3, 00:09:38.389 "num_base_bdevs_discovered": 3, 00:09:38.389 "num_base_bdevs_operational": 3, 00:09:38.389 "base_bdevs_list": [ 00:09:38.389 { 00:09:38.389 "name": "NewBaseBdev", 00:09:38.389 "uuid": "a7f3090e-09d7-494b-b316-3382977600e8", 00:09:38.389 "is_configured": true, 00:09:38.389 "data_offset": 0, 00:09:38.389 "data_size": 65536 00:09:38.389 }, 00:09:38.389 { 00:09:38.389 "name": "BaseBdev2", 00:09:38.389 "uuid": "038100a5-75b2-4d72-b1fc-7156862bb82e", 00:09:38.389 "is_configured": true, 00:09:38.389 "data_offset": 0, 00:09:38.389 "data_size": 65536 00:09:38.389 }, 00:09:38.389 { 00:09:38.389 "name": "BaseBdev3", 00:09:38.389 "uuid": "1488d52b-3f6a-49f0-b7c7-1ce28accea30", 00:09:38.389 "is_configured": true, 00:09:38.389 "data_offset": 0, 00:09:38.389 "data_size": 65536 00:09:38.389 } 00:09:38.389 ] 00:09:38.389 }' 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:38.389 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.647 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:38.647 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:38.647 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:38.647 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:38.647 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:38.647 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:38.647 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:38.647 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:38.647 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.647 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.647 [2024-11-20 15:57:36.792306] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:38.648 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.648 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:38.648 "name": "Existed_Raid", 00:09:38.648 "aliases": [ 00:09:38.648 "6bf29723-dfa2-49e1-a182-1f85b4070e1d" 00:09:38.648 ], 00:09:38.648 "product_name": "Raid Volume", 00:09:38.648 "block_size": 512, 00:09:38.648 "num_blocks": 196608, 00:09:38.648 "uuid": "6bf29723-dfa2-49e1-a182-1f85b4070e1d", 00:09:38.648 "assigned_rate_limits": { 00:09:38.648 "rw_ios_per_sec": 0, 00:09:38.648 "rw_mbytes_per_sec": 0, 00:09:38.648 "r_mbytes_per_sec": 0, 00:09:38.648 "w_mbytes_per_sec": 0 00:09:38.648 }, 00:09:38.648 "claimed": false, 00:09:38.648 "zoned": false, 00:09:38.648 "supported_io_types": { 00:09:38.648 "read": true, 00:09:38.648 "write": true, 00:09:38.648 "unmap": true, 00:09:38.648 "flush": true, 00:09:38.648 "reset": true, 00:09:38.648 "nvme_admin": false, 00:09:38.648 "nvme_io": false, 00:09:38.648 "nvme_io_md": false, 00:09:38.648 "write_zeroes": true, 00:09:38.648 "zcopy": false, 00:09:38.648 "get_zone_info": false, 00:09:38.648 "zone_management": false, 00:09:38.648 "zone_append": false, 00:09:38.648 "compare": false, 00:09:38.648 "compare_and_write": false, 00:09:38.648 "abort": false, 00:09:38.648 "seek_hole": false, 00:09:38.648 "seek_data": false, 00:09:38.648 "copy": false, 00:09:38.648 "nvme_iov_md": false 00:09:38.648 }, 00:09:38.648 "memory_domains": [ 00:09:38.648 { 00:09:38.648 "dma_device_id": "system", 00:09:38.648 "dma_device_type": 1 00:09:38.648 }, 00:09:38.648 { 00:09:38.648 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:38.648 "dma_device_type": 2 00:09:38.648 }, 00:09:38.648 { 00:09:38.648 "dma_device_id": "system", 00:09:38.648 "dma_device_type": 1 00:09:38.648 }, 00:09:38.648 { 00:09:38.648 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:38.648 "dma_device_type": 2 00:09:38.648 }, 00:09:38.648 { 00:09:38.648 "dma_device_id": "system", 00:09:38.648 "dma_device_type": 1 00:09:38.648 }, 00:09:38.648 { 00:09:38.648 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:38.648 "dma_device_type": 2 00:09:38.648 } 00:09:38.648 ], 00:09:38.648 "driver_specific": { 00:09:38.648 "raid": { 00:09:38.648 "uuid": "6bf29723-dfa2-49e1-a182-1f85b4070e1d", 00:09:38.648 "strip_size_kb": 64, 00:09:38.648 "state": "online", 00:09:38.648 "raid_level": "raid0", 00:09:38.648 "superblock": false, 00:09:38.648 "num_base_bdevs": 3, 00:09:38.648 "num_base_bdevs_discovered": 3, 00:09:38.648 "num_base_bdevs_operational": 3, 00:09:38.648 "base_bdevs_list": [ 00:09:38.648 { 00:09:38.648 "name": "NewBaseBdev", 00:09:38.648 "uuid": "a7f3090e-09d7-494b-b316-3382977600e8", 00:09:38.648 "is_configured": true, 00:09:38.648 "data_offset": 0, 00:09:38.648 "data_size": 65536 00:09:38.648 }, 00:09:38.648 { 00:09:38.648 "name": "BaseBdev2", 00:09:38.648 "uuid": "038100a5-75b2-4d72-b1fc-7156862bb82e", 00:09:38.648 "is_configured": true, 00:09:38.648 "data_offset": 0, 00:09:38.648 "data_size": 65536 00:09:38.648 }, 00:09:38.648 { 00:09:38.648 "name": "BaseBdev3", 00:09:38.648 "uuid": "1488d52b-3f6a-49f0-b7c7-1ce28accea30", 00:09:38.648 "is_configured": true, 00:09:38.648 "data_offset": 0, 00:09:38.648 "data_size": 65536 00:09:38.648 } 00:09:38.648 ] 00:09:38.648 } 00:09:38.648 } 00:09:38.648 }' 00:09:38.648 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:38.648 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:38.648 BaseBdev2 00:09:38.648 BaseBdev3' 00:09:38.648 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:38.648 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:38.648 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:38.648 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:38.648 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.648 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.648 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:38.648 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.906 [2024-11-20 15:57:36.968068] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:38.906 [2024-11-20 15:57:36.968089] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:38.906 [2024-11-20 15:57:36.968144] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:38.906 [2024-11-20 15:57:36.968191] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:38.906 [2024-11-20 15:57:36.968200] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 62328 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 62328 ']' 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 62328 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:09:38.906 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:38.907 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 62328 00:09:38.907 killing process with pid 62328 00:09:38.907 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:38.907 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:38.907 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 62328' 00:09:38.907 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 62328 00:09:38.907 [2024-11-20 15:57:36.995377] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:38.907 15:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 62328 00:09:38.907 [2024-11-20 15:57:37.140302] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:39.471 15:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:39.471 00:09:39.471 real 0m7.248s 00:09:39.471 user 0m11.711s 00:09:39.471 sys 0m1.157s 00:09:39.471 15:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:39.472 15:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.472 ************************************ 00:09:39.472 END TEST raid_state_function_test 00:09:39.472 ************************************ 00:09:39.729 15:57:37 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 3 true 00:09:39.729 15:57:37 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:39.729 15:57:37 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:39.729 15:57:37 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:39.729 ************************************ 00:09:39.729 START TEST raid_state_function_test_sb 00:09:39.729 ************************************ 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 3 true 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=62922 00:09:39.729 Process raid pid: 62922 00:09:39.729 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 62922' 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 62922 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 62922 ']' 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:39.729 15:57:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.729 [2024-11-20 15:57:37.810384] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:09:39.729 [2024-11-20 15:57:37.810493] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:39.729 [2024-11-20 15:57:37.962124] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:39.987 [2024-11-20 15:57:38.046001] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:39.987 [2024-11-20 15:57:38.155618] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:39.987 [2024-11-20 15:57:38.155651] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.553 [2024-11-20 15:57:38.654568] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:40.553 [2024-11-20 15:57:38.654609] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:40.553 [2024-11-20 15:57:38.654617] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:40.553 [2024-11-20 15:57:38.654624] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:40.553 [2024-11-20 15:57:38.654633] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:40.553 [2024-11-20 15:57:38.654640] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:40.553 "name": "Existed_Raid", 00:09:40.553 "uuid": "65888f0a-bf79-49b1-96e3-c1ff7efb65c8", 00:09:40.553 "strip_size_kb": 64, 00:09:40.553 "state": "configuring", 00:09:40.553 "raid_level": "raid0", 00:09:40.553 "superblock": true, 00:09:40.553 "num_base_bdevs": 3, 00:09:40.553 "num_base_bdevs_discovered": 0, 00:09:40.553 "num_base_bdevs_operational": 3, 00:09:40.553 "base_bdevs_list": [ 00:09:40.553 { 00:09:40.553 "name": "BaseBdev1", 00:09:40.553 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:40.553 "is_configured": false, 00:09:40.553 "data_offset": 0, 00:09:40.553 "data_size": 0 00:09:40.553 }, 00:09:40.553 { 00:09:40.553 "name": "BaseBdev2", 00:09:40.553 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:40.553 "is_configured": false, 00:09:40.553 "data_offset": 0, 00:09:40.553 "data_size": 0 00:09:40.553 }, 00:09:40.553 { 00:09:40.553 "name": "BaseBdev3", 00:09:40.553 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:40.553 "is_configured": false, 00:09:40.553 "data_offset": 0, 00:09:40.553 "data_size": 0 00:09:40.553 } 00:09:40.553 ] 00:09:40.553 }' 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:40.553 15:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.811 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:40.811 15:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.811 15:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.812 [2024-11-20 15:57:38.970577] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:40.812 [2024-11-20 15:57:38.970603] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:09:40.812 15:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.812 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:40.812 15:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.812 15:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.812 [2024-11-20 15:57:38.978590] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:40.812 [2024-11-20 15:57:38.978625] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:40.812 [2024-11-20 15:57:38.978632] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:40.812 [2024-11-20 15:57:38.978640] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:40.812 [2024-11-20 15:57:38.978645] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:40.812 [2024-11-20 15:57:38.978652] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:40.812 15:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.812 15:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:40.812 15:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.812 15:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.812 [2024-11-20 15:57:39.006721] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:40.812 BaseBdev1 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.812 [ 00:09:40.812 { 00:09:40.812 "name": "BaseBdev1", 00:09:40.812 "aliases": [ 00:09:40.812 "05f9ea37-62db-4195-9f97-c778c2834afc" 00:09:40.812 ], 00:09:40.812 "product_name": "Malloc disk", 00:09:40.812 "block_size": 512, 00:09:40.812 "num_blocks": 65536, 00:09:40.812 "uuid": "05f9ea37-62db-4195-9f97-c778c2834afc", 00:09:40.812 "assigned_rate_limits": { 00:09:40.812 "rw_ios_per_sec": 0, 00:09:40.812 "rw_mbytes_per_sec": 0, 00:09:40.812 "r_mbytes_per_sec": 0, 00:09:40.812 "w_mbytes_per_sec": 0 00:09:40.812 }, 00:09:40.812 "claimed": true, 00:09:40.812 "claim_type": "exclusive_write", 00:09:40.812 "zoned": false, 00:09:40.812 "supported_io_types": { 00:09:40.812 "read": true, 00:09:40.812 "write": true, 00:09:40.812 "unmap": true, 00:09:40.812 "flush": true, 00:09:40.812 "reset": true, 00:09:40.812 "nvme_admin": false, 00:09:40.812 "nvme_io": false, 00:09:40.812 "nvme_io_md": false, 00:09:40.812 "write_zeroes": true, 00:09:40.812 "zcopy": true, 00:09:40.812 "get_zone_info": false, 00:09:40.812 "zone_management": false, 00:09:40.812 "zone_append": false, 00:09:40.812 "compare": false, 00:09:40.812 "compare_and_write": false, 00:09:40.812 "abort": true, 00:09:40.812 "seek_hole": false, 00:09:40.812 "seek_data": false, 00:09:40.812 "copy": true, 00:09:40.812 "nvme_iov_md": false 00:09:40.812 }, 00:09:40.812 "memory_domains": [ 00:09:40.812 { 00:09:40.812 "dma_device_id": "system", 00:09:40.812 "dma_device_type": 1 00:09:40.812 }, 00:09:40.812 { 00:09:40.812 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:40.812 "dma_device_type": 2 00:09:40.812 } 00:09:40.812 ], 00:09:40.812 "driver_specific": {} 00:09:40.812 } 00:09:40.812 ] 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.812 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.069 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:41.069 "name": "Existed_Raid", 00:09:41.069 "uuid": "ea6d133b-dc1d-4339-97f2-e7c82731f46f", 00:09:41.069 "strip_size_kb": 64, 00:09:41.069 "state": "configuring", 00:09:41.069 "raid_level": "raid0", 00:09:41.069 "superblock": true, 00:09:41.069 "num_base_bdevs": 3, 00:09:41.069 "num_base_bdevs_discovered": 1, 00:09:41.069 "num_base_bdevs_operational": 3, 00:09:41.069 "base_bdevs_list": [ 00:09:41.069 { 00:09:41.069 "name": "BaseBdev1", 00:09:41.069 "uuid": "05f9ea37-62db-4195-9f97-c778c2834afc", 00:09:41.069 "is_configured": true, 00:09:41.069 "data_offset": 2048, 00:09:41.069 "data_size": 63488 00:09:41.069 }, 00:09:41.069 { 00:09:41.069 "name": "BaseBdev2", 00:09:41.069 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:41.069 "is_configured": false, 00:09:41.069 "data_offset": 0, 00:09:41.069 "data_size": 0 00:09:41.069 }, 00:09:41.069 { 00:09:41.069 "name": "BaseBdev3", 00:09:41.069 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:41.069 "is_configured": false, 00:09:41.069 "data_offset": 0, 00:09:41.069 "data_size": 0 00:09:41.069 } 00:09:41.069 ] 00:09:41.069 }' 00:09:41.069 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:41.069 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.327 [2024-11-20 15:57:39.354812] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:41.327 [2024-11-20 15:57:39.354852] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.327 [2024-11-20 15:57:39.362864] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:41.327 [2024-11-20 15:57:39.364387] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:41.327 [2024-11-20 15:57:39.364425] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:41.327 [2024-11-20 15:57:39.364432] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:41.327 [2024-11-20 15:57:39.364439] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:41.327 "name": "Existed_Raid", 00:09:41.327 "uuid": "15e92d0f-3079-4f3c-9532-35371c186772", 00:09:41.327 "strip_size_kb": 64, 00:09:41.327 "state": "configuring", 00:09:41.327 "raid_level": "raid0", 00:09:41.327 "superblock": true, 00:09:41.327 "num_base_bdevs": 3, 00:09:41.327 "num_base_bdevs_discovered": 1, 00:09:41.327 "num_base_bdevs_operational": 3, 00:09:41.327 "base_bdevs_list": [ 00:09:41.327 { 00:09:41.327 "name": "BaseBdev1", 00:09:41.327 "uuid": "05f9ea37-62db-4195-9f97-c778c2834afc", 00:09:41.327 "is_configured": true, 00:09:41.327 "data_offset": 2048, 00:09:41.327 "data_size": 63488 00:09:41.327 }, 00:09:41.327 { 00:09:41.327 "name": "BaseBdev2", 00:09:41.327 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:41.327 "is_configured": false, 00:09:41.327 "data_offset": 0, 00:09:41.327 "data_size": 0 00:09:41.327 }, 00:09:41.327 { 00:09:41.327 "name": "BaseBdev3", 00:09:41.327 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:41.327 "is_configured": false, 00:09:41.327 "data_offset": 0, 00:09:41.327 "data_size": 0 00:09:41.327 } 00:09:41.327 ] 00:09:41.327 }' 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:41.327 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.584 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.585 [2024-11-20 15:57:39.692969] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:41.585 BaseBdev2 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.585 [ 00:09:41.585 { 00:09:41.585 "name": "BaseBdev2", 00:09:41.585 "aliases": [ 00:09:41.585 "d31f78c7-f520-48db-975e-d6f26998e1a9" 00:09:41.585 ], 00:09:41.585 "product_name": "Malloc disk", 00:09:41.585 "block_size": 512, 00:09:41.585 "num_blocks": 65536, 00:09:41.585 "uuid": "d31f78c7-f520-48db-975e-d6f26998e1a9", 00:09:41.585 "assigned_rate_limits": { 00:09:41.585 "rw_ios_per_sec": 0, 00:09:41.585 "rw_mbytes_per_sec": 0, 00:09:41.585 "r_mbytes_per_sec": 0, 00:09:41.585 "w_mbytes_per_sec": 0 00:09:41.585 }, 00:09:41.585 "claimed": true, 00:09:41.585 "claim_type": "exclusive_write", 00:09:41.585 "zoned": false, 00:09:41.585 "supported_io_types": { 00:09:41.585 "read": true, 00:09:41.585 "write": true, 00:09:41.585 "unmap": true, 00:09:41.585 "flush": true, 00:09:41.585 "reset": true, 00:09:41.585 "nvme_admin": false, 00:09:41.585 "nvme_io": false, 00:09:41.585 "nvme_io_md": false, 00:09:41.585 "write_zeroes": true, 00:09:41.585 "zcopy": true, 00:09:41.585 "get_zone_info": false, 00:09:41.585 "zone_management": false, 00:09:41.585 "zone_append": false, 00:09:41.585 "compare": false, 00:09:41.585 "compare_and_write": false, 00:09:41.585 "abort": true, 00:09:41.585 "seek_hole": false, 00:09:41.585 "seek_data": false, 00:09:41.585 "copy": true, 00:09:41.585 "nvme_iov_md": false 00:09:41.585 }, 00:09:41.585 "memory_domains": [ 00:09:41.585 { 00:09:41.585 "dma_device_id": "system", 00:09:41.585 "dma_device_type": 1 00:09:41.585 }, 00:09:41.585 { 00:09:41.585 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:41.585 "dma_device_type": 2 00:09:41.585 } 00:09:41.585 ], 00:09:41.585 "driver_specific": {} 00:09:41.585 } 00:09:41.585 ] 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:41.585 "name": "Existed_Raid", 00:09:41.585 "uuid": "15e92d0f-3079-4f3c-9532-35371c186772", 00:09:41.585 "strip_size_kb": 64, 00:09:41.585 "state": "configuring", 00:09:41.585 "raid_level": "raid0", 00:09:41.585 "superblock": true, 00:09:41.585 "num_base_bdevs": 3, 00:09:41.585 "num_base_bdevs_discovered": 2, 00:09:41.585 "num_base_bdevs_operational": 3, 00:09:41.585 "base_bdevs_list": [ 00:09:41.585 { 00:09:41.585 "name": "BaseBdev1", 00:09:41.585 "uuid": "05f9ea37-62db-4195-9f97-c778c2834afc", 00:09:41.585 "is_configured": true, 00:09:41.585 "data_offset": 2048, 00:09:41.585 "data_size": 63488 00:09:41.585 }, 00:09:41.585 { 00:09:41.585 "name": "BaseBdev2", 00:09:41.585 "uuid": "d31f78c7-f520-48db-975e-d6f26998e1a9", 00:09:41.585 "is_configured": true, 00:09:41.585 "data_offset": 2048, 00:09:41.585 "data_size": 63488 00:09:41.585 }, 00:09:41.585 { 00:09:41.585 "name": "BaseBdev3", 00:09:41.585 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:41.585 "is_configured": false, 00:09:41.585 "data_offset": 0, 00:09:41.585 "data_size": 0 00:09:41.585 } 00:09:41.585 ] 00:09:41.585 }' 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:41.585 15:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.843 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:41.843 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.843 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.843 [2024-11-20 15:57:40.060453] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:41.843 [2024-11-20 15:57:40.060849] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:41.843 [2024-11-20 15:57:40.060978] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:41.843 BaseBdev3 00:09:41.843 [2024-11-20 15:57:40.061216] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:09:41.843 [2024-11-20 15:57:40.061393] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:41.843 [2024-11-20 15:57:40.061451] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:09:41.843 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.843 [2024-11-20 15:57:40.061631] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:41.843 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:41.843 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:41.843 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:41.843 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:41.843 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:41.843 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:41.843 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.844 [ 00:09:41.844 { 00:09:41.844 "name": "BaseBdev3", 00:09:41.844 "aliases": [ 00:09:41.844 "80fa1742-14b8-49d0-80de-57c40716a9c7" 00:09:41.844 ], 00:09:41.844 "product_name": "Malloc disk", 00:09:41.844 "block_size": 512, 00:09:41.844 "num_blocks": 65536, 00:09:41.844 "uuid": "80fa1742-14b8-49d0-80de-57c40716a9c7", 00:09:41.844 "assigned_rate_limits": { 00:09:41.844 "rw_ios_per_sec": 0, 00:09:41.844 "rw_mbytes_per_sec": 0, 00:09:41.844 "r_mbytes_per_sec": 0, 00:09:41.844 "w_mbytes_per_sec": 0 00:09:41.844 }, 00:09:41.844 "claimed": true, 00:09:41.844 "claim_type": "exclusive_write", 00:09:41.844 "zoned": false, 00:09:41.844 "supported_io_types": { 00:09:41.844 "read": true, 00:09:41.844 "write": true, 00:09:41.844 "unmap": true, 00:09:41.844 "flush": true, 00:09:41.844 "reset": true, 00:09:41.844 "nvme_admin": false, 00:09:41.844 "nvme_io": false, 00:09:41.844 "nvme_io_md": false, 00:09:41.844 "write_zeroes": true, 00:09:41.844 "zcopy": true, 00:09:41.844 "get_zone_info": false, 00:09:41.844 "zone_management": false, 00:09:41.844 "zone_append": false, 00:09:41.844 "compare": false, 00:09:41.844 "compare_and_write": false, 00:09:41.844 "abort": true, 00:09:41.844 "seek_hole": false, 00:09:41.844 "seek_data": false, 00:09:41.844 "copy": true, 00:09:41.844 "nvme_iov_md": false 00:09:41.844 }, 00:09:41.844 "memory_domains": [ 00:09:41.844 { 00:09:41.844 "dma_device_id": "system", 00:09:41.844 "dma_device_type": 1 00:09:41.844 }, 00:09:41.844 { 00:09:41.844 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:41.844 "dma_device_type": 2 00:09:41.844 } 00:09:41.844 ], 00:09:41.844 "driver_specific": {} 00:09:41.844 } 00:09:41.844 ] 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.844 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.100 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:42.100 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.100 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:42.100 "name": "Existed_Raid", 00:09:42.100 "uuid": "15e92d0f-3079-4f3c-9532-35371c186772", 00:09:42.100 "strip_size_kb": 64, 00:09:42.100 "state": "online", 00:09:42.100 "raid_level": "raid0", 00:09:42.100 "superblock": true, 00:09:42.100 "num_base_bdevs": 3, 00:09:42.100 "num_base_bdevs_discovered": 3, 00:09:42.100 "num_base_bdevs_operational": 3, 00:09:42.100 "base_bdevs_list": [ 00:09:42.100 { 00:09:42.100 "name": "BaseBdev1", 00:09:42.100 "uuid": "05f9ea37-62db-4195-9f97-c778c2834afc", 00:09:42.100 "is_configured": true, 00:09:42.100 "data_offset": 2048, 00:09:42.100 "data_size": 63488 00:09:42.100 }, 00:09:42.100 { 00:09:42.100 "name": "BaseBdev2", 00:09:42.100 "uuid": "d31f78c7-f520-48db-975e-d6f26998e1a9", 00:09:42.100 "is_configured": true, 00:09:42.100 "data_offset": 2048, 00:09:42.100 "data_size": 63488 00:09:42.100 }, 00:09:42.100 { 00:09:42.100 "name": "BaseBdev3", 00:09:42.100 "uuid": "80fa1742-14b8-49d0-80de-57c40716a9c7", 00:09:42.100 "is_configured": true, 00:09:42.100 "data_offset": 2048, 00:09:42.100 "data_size": 63488 00:09:42.100 } 00:09:42.100 ] 00:09:42.100 }' 00:09:42.100 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:42.100 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.357 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:42.357 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:42.357 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:42.357 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:42.357 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:42.357 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:42.357 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:42.357 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:42.357 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.357 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.357 [2024-11-20 15:57:40.416834] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:42.357 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.357 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:42.357 "name": "Existed_Raid", 00:09:42.357 "aliases": [ 00:09:42.357 "15e92d0f-3079-4f3c-9532-35371c186772" 00:09:42.357 ], 00:09:42.357 "product_name": "Raid Volume", 00:09:42.357 "block_size": 512, 00:09:42.357 "num_blocks": 190464, 00:09:42.357 "uuid": "15e92d0f-3079-4f3c-9532-35371c186772", 00:09:42.357 "assigned_rate_limits": { 00:09:42.357 "rw_ios_per_sec": 0, 00:09:42.357 "rw_mbytes_per_sec": 0, 00:09:42.357 "r_mbytes_per_sec": 0, 00:09:42.357 "w_mbytes_per_sec": 0 00:09:42.357 }, 00:09:42.357 "claimed": false, 00:09:42.357 "zoned": false, 00:09:42.357 "supported_io_types": { 00:09:42.357 "read": true, 00:09:42.357 "write": true, 00:09:42.357 "unmap": true, 00:09:42.357 "flush": true, 00:09:42.357 "reset": true, 00:09:42.357 "nvme_admin": false, 00:09:42.357 "nvme_io": false, 00:09:42.357 "nvme_io_md": false, 00:09:42.357 "write_zeroes": true, 00:09:42.357 "zcopy": false, 00:09:42.357 "get_zone_info": false, 00:09:42.357 "zone_management": false, 00:09:42.357 "zone_append": false, 00:09:42.357 "compare": false, 00:09:42.357 "compare_and_write": false, 00:09:42.357 "abort": false, 00:09:42.357 "seek_hole": false, 00:09:42.357 "seek_data": false, 00:09:42.357 "copy": false, 00:09:42.357 "nvme_iov_md": false 00:09:42.357 }, 00:09:42.357 "memory_domains": [ 00:09:42.357 { 00:09:42.357 "dma_device_id": "system", 00:09:42.357 "dma_device_type": 1 00:09:42.357 }, 00:09:42.357 { 00:09:42.357 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:42.357 "dma_device_type": 2 00:09:42.357 }, 00:09:42.357 { 00:09:42.357 "dma_device_id": "system", 00:09:42.357 "dma_device_type": 1 00:09:42.357 }, 00:09:42.357 { 00:09:42.357 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:42.357 "dma_device_type": 2 00:09:42.357 }, 00:09:42.357 { 00:09:42.357 "dma_device_id": "system", 00:09:42.357 "dma_device_type": 1 00:09:42.357 }, 00:09:42.357 { 00:09:42.357 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:42.357 "dma_device_type": 2 00:09:42.357 } 00:09:42.357 ], 00:09:42.357 "driver_specific": { 00:09:42.357 "raid": { 00:09:42.357 "uuid": "15e92d0f-3079-4f3c-9532-35371c186772", 00:09:42.357 "strip_size_kb": 64, 00:09:42.357 "state": "online", 00:09:42.357 "raid_level": "raid0", 00:09:42.357 "superblock": true, 00:09:42.357 "num_base_bdevs": 3, 00:09:42.357 "num_base_bdevs_discovered": 3, 00:09:42.357 "num_base_bdevs_operational": 3, 00:09:42.357 "base_bdevs_list": [ 00:09:42.357 { 00:09:42.357 "name": "BaseBdev1", 00:09:42.357 "uuid": "05f9ea37-62db-4195-9f97-c778c2834afc", 00:09:42.357 "is_configured": true, 00:09:42.357 "data_offset": 2048, 00:09:42.357 "data_size": 63488 00:09:42.357 }, 00:09:42.357 { 00:09:42.357 "name": "BaseBdev2", 00:09:42.357 "uuid": "d31f78c7-f520-48db-975e-d6f26998e1a9", 00:09:42.357 "is_configured": true, 00:09:42.357 "data_offset": 2048, 00:09:42.357 "data_size": 63488 00:09:42.357 }, 00:09:42.357 { 00:09:42.357 "name": "BaseBdev3", 00:09:42.357 "uuid": "80fa1742-14b8-49d0-80de-57c40716a9c7", 00:09:42.357 "is_configured": true, 00:09:42.357 "data_offset": 2048, 00:09:42.357 "data_size": 63488 00:09:42.357 } 00:09:42.357 ] 00:09:42.357 } 00:09:42.357 } 00:09:42.357 }' 00:09:42.357 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:42.357 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:42.358 BaseBdev2 00:09:42.358 BaseBdev3' 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.358 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.358 [2024-11-20 15:57:40.600629] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:42.358 [2024-11-20 15:57:40.600755] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:42.358 [2024-11-20 15:57:40.600809] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:42.615 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.615 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:42.615 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:09:42.615 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:42.615 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:09:42.615 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:42.615 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:09:42.616 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:42.616 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:42.616 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:42.616 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:42.616 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:42.616 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:42.616 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:42.616 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:42.616 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:42.616 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:42.616 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:42.616 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.616 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.616 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.616 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:42.616 "name": "Existed_Raid", 00:09:42.616 "uuid": "15e92d0f-3079-4f3c-9532-35371c186772", 00:09:42.616 "strip_size_kb": 64, 00:09:42.616 "state": "offline", 00:09:42.616 "raid_level": "raid0", 00:09:42.616 "superblock": true, 00:09:42.616 "num_base_bdevs": 3, 00:09:42.616 "num_base_bdevs_discovered": 2, 00:09:42.616 "num_base_bdevs_operational": 2, 00:09:42.616 "base_bdevs_list": [ 00:09:42.616 { 00:09:42.616 "name": null, 00:09:42.616 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:42.616 "is_configured": false, 00:09:42.616 "data_offset": 0, 00:09:42.616 "data_size": 63488 00:09:42.616 }, 00:09:42.616 { 00:09:42.616 "name": "BaseBdev2", 00:09:42.616 "uuid": "d31f78c7-f520-48db-975e-d6f26998e1a9", 00:09:42.616 "is_configured": true, 00:09:42.616 "data_offset": 2048, 00:09:42.616 "data_size": 63488 00:09:42.616 }, 00:09:42.616 { 00:09:42.616 "name": "BaseBdev3", 00:09:42.616 "uuid": "80fa1742-14b8-49d0-80de-57c40716a9c7", 00:09:42.616 "is_configured": true, 00:09:42.616 "data_offset": 2048, 00:09:42.616 "data_size": 63488 00:09:42.616 } 00:09:42.616 ] 00:09:42.616 }' 00:09:42.616 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:42.616 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.873 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:42.873 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:42.873 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:42.873 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:42.873 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.873 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.874 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.874 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:42.874 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:42.874 15:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:42.874 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.874 15:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.874 [2024-11-20 15:57:40.983923] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:42.874 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.874 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:42.874 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:42.874 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:42.874 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.874 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.874 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:42.874 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.874 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:42.874 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:42.874 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:42.874 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.874 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.874 [2024-11-20 15:57:41.074585] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:42.874 [2024-11-20 15:57:41.074625] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:09:42.874 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.131 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:43.131 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:43.131 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.131 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.131 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.131 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:43.131 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.131 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:43.131 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:43.131 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:43.131 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:43.131 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:43.131 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:43.131 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.132 BaseBdev2 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.132 [ 00:09:43.132 { 00:09:43.132 "name": "BaseBdev2", 00:09:43.132 "aliases": [ 00:09:43.132 "f1bafcc1-228d-498c-8b2e-a9af417030e4" 00:09:43.132 ], 00:09:43.132 "product_name": "Malloc disk", 00:09:43.132 "block_size": 512, 00:09:43.132 "num_blocks": 65536, 00:09:43.132 "uuid": "f1bafcc1-228d-498c-8b2e-a9af417030e4", 00:09:43.132 "assigned_rate_limits": { 00:09:43.132 "rw_ios_per_sec": 0, 00:09:43.132 "rw_mbytes_per_sec": 0, 00:09:43.132 "r_mbytes_per_sec": 0, 00:09:43.132 "w_mbytes_per_sec": 0 00:09:43.132 }, 00:09:43.132 "claimed": false, 00:09:43.132 "zoned": false, 00:09:43.132 "supported_io_types": { 00:09:43.132 "read": true, 00:09:43.132 "write": true, 00:09:43.132 "unmap": true, 00:09:43.132 "flush": true, 00:09:43.132 "reset": true, 00:09:43.132 "nvme_admin": false, 00:09:43.132 "nvme_io": false, 00:09:43.132 "nvme_io_md": false, 00:09:43.132 "write_zeroes": true, 00:09:43.132 "zcopy": true, 00:09:43.132 "get_zone_info": false, 00:09:43.132 "zone_management": false, 00:09:43.132 "zone_append": false, 00:09:43.132 "compare": false, 00:09:43.132 "compare_and_write": false, 00:09:43.132 "abort": true, 00:09:43.132 "seek_hole": false, 00:09:43.132 "seek_data": false, 00:09:43.132 "copy": true, 00:09:43.132 "nvme_iov_md": false 00:09:43.132 }, 00:09:43.132 "memory_domains": [ 00:09:43.132 { 00:09:43.132 "dma_device_id": "system", 00:09:43.132 "dma_device_type": 1 00:09:43.132 }, 00:09:43.132 { 00:09:43.132 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:43.132 "dma_device_type": 2 00:09:43.132 } 00:09:43.132 ], 00:09:43.132 "driver_specific": {} 00:09:43.132 } 00:09:43.132 ] 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.132 BaseBdev3 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.132 [ 00:09:43.132 { 00:09:43.132 "name": "BaseBdev3", 00:09:43.132 "aliases": [ 00:09:43.132 "cf93741b-f85f-4200-b240-a2d3937d1479" 00:09:43.132 ], 00:09:43.132 "product_name": "Malloc disk", 00:09:43.132 "block_size": 512, 00:09:43.132 "num_blocks": 65536, 00:09:43.132 "uuid": "cf93741b-f85f-4200-b240-a2d3937d1479", 00:09:43.132 "assigned_rate_limits": { 00:09:43.132 "rw_ios_per_sec": 0, 00:09:43.132 "rw_mbytes_per_sec": 0, 00:09:43.132 "r_mbytes_per_sec": 0, 00:09:43.132 "w_mbytes_per_sec": 0 00:09:43.132 }, 00:09:43.132 "claimed": false, 00:09:43.132 "zoned": false, 00:09:43.132 "supported_io_types": { 00:09:43.132 "read": true, 00:09:43.132 "write": true, 00:09:43.132 "unmap": true, 00:09:43.132 "flush": true, 00:09:43.132 "reset": true, 00:09:43.132 "nvme_admin": false, 00:09:43.132 "nvme_io": false, 00:09:43.132 "nvme_io_md": false, 00:09:43.132 "write_zeroes": true, 00:09:43.132 "zcopy": true, 00:09:43.132 "get_zone_info": false, 00:09:43.132 "zone_management": false, 00:09:43.132 "zone_append": false, 00:09:43.132 "compare": false, 00:09:43.132 "compare_and_write": false, 00:09:43.132 "abort": true, 00:09:43.132 "seek_hole": false, 00:09:43.132 "seek_data": false, 00:09:43.132 "copy": true, 00:09:43.132 "nvme_iov_md": false 00:09:43.132 }, 00:09:43.132 "memory_domains": [ 00:09:43.132 { 00:09:43.132 "dma_device_id": "system", 00:09:43.132 "dma_device_type": 1 00:09:43.132 }, 00:09:43.132 { 00:09:43.132 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:43.132 "dma_device_type": 2 00:09:43.132 } 00:09:43.132 ], 00:09:43.132 "driver_specific": {} 00:09:43.132 } 00:09:43.132 ] 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.132 [2024-11-20 15:57:41.273159] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:43.132 [2024-11-20 15:57:41.273310] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:43.132 [2024-11-20 15:57:41.273375] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:43.132 [2024-11-20 15:57:41.274958] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.132 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:43.132 "name": "Existed_Raid", 00:09:43.132 "uuid": "2db3b179-1bf8-4b92-b22c-493cbaf50220", 00:09:43.132 "strip_size_kb": 64, 00:09:43.132 "state": "configuring", 00:09:43.132 "raid_level": "raid0", 00:09:43.132 "superblock": true, 00:09:43.132 "num_base_bdevs": 3, 00:09:43.132 "num_base_bdevs_discovered": 2, 00:09:43.132 "num_base_bdevs_operational": 3, 00:09:43.133 "base_bdevs_list": [ 00:09:43.133 { 00:09:43.133 "name": "BaseBdev1", 00:09:43.133 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:43.133 "is_configured": false, 00:09:43.133 "data_offset": 0, 00:09:43.133 "data_size": 0 00:09:43.133 }, 00:09:43.133 { 00:09:43.133 "name": "BaseBdev2", 00:09:43.133 "uuid": "f1bafcc1-228d-498c-8b2e-a9af417030e4", 00:09:43.133 "is_configured": true, 00:09:43.133 "data_offset": 2048, 00:09:43.133 "data_size": 63488 00:09:43.133 }, 00:09:43.133 { 00:09:43.133 "name": "BaseBdev3", 00:09:43.133 "uuid": "cf93741b-f85f-4200-b240-a2d3937d1479", 00:09:43.133 "is_configured": true, 00:09:43.133 "data_offset": 2048, 00:09:43.133 "data_size": 63488 00:09:43.133 } 00:09:43.133 ] 00:09:43.133 }' 00:09:43.133 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:43.133 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.390 [2024-11-20 15:57:41.597212] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:43.390 "name": "Existed_Raid", 00:09:43.390 "uuid": "2db3b179-1bf8-4b92-b22c-493cbaf50220", 00:09:43.390 "strip_size_kb": 64, 00:09:43.390 "state": "configuring", 00:09:43.390 "raid_level": "raid0", 00:09:43.390 "superblock": true, 00:09:43.390 "num_base_bdevs": 3, 00:09:43.390 "num_base_bdevs_discovered": 1, 00:09:43.390 "num_base_bdevs_operational": 3, 00:09:43.390 "base_bdevs_list": [ 00:09:43.390 { 00:09:43.390 "name": "BaseBdev1", 00:09:43.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:43.390 "is_configured": false, 00:09:43.390 "data_offset": 0, 00:09:43.390 "data_size": 0 00:09:43.390 }, 00:09:43.390 { 00:09:43.390 "name": null, 00:09:43.390 "uuid": "f1bafcc1-228d-498c-8b2e-a9af417030e4", 00:09:43.390 "is_configured": false, 00:09:43.390 "data_offset": 0, 00:09:43.390 "data_size": 63488 00:09:43.390 }, 00:09:43.390 { 00:09:43.390 "name": "BaseBdev3", 00:09:43.390 "uuid": "cf93741b-f85f-4200-b240-a2d3937d1479", 00:09:43.390 "is_configured": true, 00:09:43.390 "data_offset": 2048, 00:09:43.390 "data_size": 63488 00:09:43.390 } 00:09:43.390 ] 00:09:43.390 }' 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:43.390 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.954 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.954 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:43.954 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.954 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.954 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.954 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:43.954 15:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:43.954 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.954 15:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.954 [2024-11-20 15:57:42.003971] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:43.954 BaseBdev1 00:09:43.954 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.954 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:43.954 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:43.954 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:43.954 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:43.954 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:43.954 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:43.954 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:43.954 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.954 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.954 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.954 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:43.954 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.954 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.954 [ 00:09:43.954 { 00:09:43.954 "name": "BaseBdev1", 00:09:43.954 "aliases": [ 00:09:43.954 "9131a959-6e0a-4dc6-a386-ed259b3a7a20" 00:09:43.954 ], 00:09:43.954 "product_name": "Malloc disk", 00:09:43.954 "block_size": 512, 00:09:43.954 "num_blocks": 65536, 00:09:43.954 "uuid": "9131a959-6e0a-4dc6-a386-ed259b3a7a20", 00:09:43.954 "assigned_rate_limits": { 00:09:43.954 "rw_ios_per_sec": 0, 00:09:43.954 "rw_mbytes_per_sec": 0, 00:09:43.954 "r_mbytes_per_sec": 0, 00:09:43.954 "w_mbytes_per_sec": 0 00:09:43.954 }, 00:09:43.954 "claimed": true, 00:09:43.954 "claim_type": "exclusive_write", 00:09:43.954 "zoned": false, 00:09:43.954 "supported_io_types": { 00:09:43.954 "read": true, 00:09:43.954 "write": true, 00:09:43.954 "unmap": true, 00:09:43.954 "flush": true, 00:09:43.954 "reset": true, 00:09:43.954 "nvme_admin": false, 00:09:43.954 "nvme_io": false, 00:09:43.954 "nvme_io_md": false, 00:09:43.954 "write_zeroes": true, 00:09:43.954 "zcopy": true, 00:09:43.954 "get_zone_info": false, 00:09:43.954 "zone_management": false, 00:09:43.954 "zone_append": false, 00:09:43.954 "compare": false, 00:09:43.954 "compare_and_write": false, 00:09:43.954 "abort": true, 00:09:43.954 "seek_hole": false, 00:09:43.954 "seek_data": false, 00:09:43.954 "copy": true, 00:09:43.954 "nvme_iov_md": false 00:09:43.954 }, 00:09:43.954 "memory_domains": [ 00:09:43.954 { 00:09:43.955 "dma_device_id": "system", 00:09:43.955 "dma_device_type": 1 00:09:43.955 }, 00:09:43.955 { 00:09:43.955 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:43.955 "dma_device_type": 2 00:09:43.955 } 00:09:43.955 ], 00:09:43.955 "driver_specific": {} 00:09:43.955 } 00:09:43.955 ] 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:43.955 "name": "Existed_Raid", 00:09:43.955 "uuid": "2db3b179-1bf8-4b92-b22c-493cbaf50220", 00:09:43.955 "strip_size_kb": 64, 00:09:43.955 "state": "configuring", 00:09:43.955 "raid_level": "raid0", 00:09:43.955 "superblock": true, 00:09:43.955 "num_base_bdevs": 3, 00:09:43.955 "num_base_bdevs_discovered": 2, 00:09:43.955 "num_base_bdevs_operational": 3, 00:09:43.955 "base_bdevs_list": [ 00:09:43.955 { 00:09:43.955 "name": "BaseBdev1", 00:09:43.955 "uuid": "9131a959-6e0a-4dc6-a386-ed259b3a7a20", 00:09:43.955 "is_configured": true, 00:09:43.955 "data_offset": 2048, 00:09:43.955 "data_size": 63488 00:09:43.955 }, 00:09:43.955 { 00:09:43.955 "name": null, 00:09:43.955 "uuid": "f1bafcc1-228d-498c-8b2e-a9af417030e4", 00:09:43.955 "is_configured": false, 00:09:43.955 "data_offset": 0, 00:09:43.955 "data_size": 63488 00:09:43.955 }, 00:09:43.955 { 00:09:43.955 "name": "BaseBdev3", 00:09:43.955 "uuid": "cf93741b-f85f-4200-b240-a2d3937d1479", 00:09:43.955 "is_configured": true, 00:09:43.955 "data_offset": 2048, 00:09:43.955 "data_size": 63488 00:09:43.955 } 00:09:43.955 ] 00:09:43.955 }' 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:43.955 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:44.212 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:44.213 [2024-11-20 15:57:42.404093] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:44.213 "name": "Existed_Raid", 00:09:44.213 "uuid": "2db3b179-1bf8-4b92-b22c-493cbaf50220", 00:09:44.213 "strip_size_kb": 64, 00:09:44.213 "state": "configuring", 00:09:44.213 "raid_level": "raid0", 00:09:44.213 "superblock": true, 00:09:44.213 "num_base_bdevs": 3, 00:09:44.213 "num_base_bdevs_discovered": 1, 00:09:44.213 "num_base_bdevs_operational": 3, 00:09:44.213 "base_bdevs_list": [ 00:09:44.213 { 00:09:44.213 "name": "BaseBdev1", 00:09:44.213 "uuid": "9131a959-6e0a-4dc6-a386-ed259b3a7a20", 00:09:44.213 "is_configured": true, 00:09:44.213 "data_offset": 2048, 00:09:44.213 "data_size": 63488 00:09:44.213 }, 00:09:44.213 { 00:09:44.213 "name": null, 00:09:44.213 "uuid": "f1bafcc1-228d-498c-8b2e-a9af417030e4", 00:09:44.213 "is_configured": false, 00:09:44.213 "data_offset": 0, 00:09:44.213 "data_size": 63488 00:09:44.213 }, 00:09:44.213 { 00:09:44.213 "name": null, 00:09:44.213 "uuid": "cf93741b-f85f-4200-b240-a2d3937d1479", 00:09:44.213 "is_configured": false, 00:09:44.213 "data_offset": 0, 00:09:44.213 "data_size": 63488 00:09:44.213 } 00:09:44.213 ] 00:09:44.213 }' 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:44.213 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:44.472 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:44.472 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.472 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.472 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:44.731 [2024-11-20 15:57:42.748180] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:44.731 "name": "Existed_Raid", 00:09:44.731 "uuid": "2db3b179-1bf8-4b92-b22c-493cbaf50220", 00:09:44.731 "strip_size_kb": 64, 00:09:44.731 "state": "configuring", 00:09:44.731 "raid_level": "raid0", 00:09:44.731 "superblock": true, 00:09:44.731 "num_base_bdevs": 3, 00:09:44.731 "num_base_bdevs_discovered": 2, 00:09:44.731 "num_base_bdevs_operational": 3, 00:09:44.731 "base_bdevs_list": [ 00:09:44.731 { 00:09:44.731 "name": "BaseBdev1", 00:09:44.731 "uuid": "9131a959-6e0a-4dc6-a386-ed259b3a7a20", 00:09:44.731 "is_configured": true, 00:09:44.731 "data_offset": 2048, 00:09:44.731 "data_size": 63488 00:09:44.731 }, 00:09:44.731 { 00:09:44.731 "name": null, 00:09:44.731 "uuid": "f1bafcc1-228d-498c-8b2e-a9af417030e4", 00:09:44.731 "is_configured": false, 00:09:44.731 "data_offset": 0, 00:09:44.731 "data_size": 63488 00:09:44.731 }, 00:09:44.731 { 00:09:44.731 "name": "BaseBdev3", 00:09:44.731 "uuid": "cf93741b-f85f-4200-b240-a2d3937d1479", 00:09:44.731 "is_configured": true, 00:09:44.731 "data_offset": 2048, 00:09:44.731 "data_size": 63488 00:09:44.731 } 00:09:44.731 ] 00:09:44.731 }' 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:44.731 15:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:44.989 [2024-11-20 15:57:43.132261] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.989 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:44.990 "name": "Existed_Raid", 00:09:44.990 "uuid": "2db3b179-1bf8-4b92-b22c-493cbaf50220", 00:09:44.990 "strip_size_kb": 64, 00:09:44.990 "state": "configuring", 00:09:44.990 "raid_level": "raid0", 00:09:44.990 "superblock": true, 00:09:44.990 "num_base_bdevs": 3, 00:09:44.990 "num_base_bdevs_discovered": 1, 00:09:44.990 "num_base_bdevs_operational": 3, 00:09:44.990 "base_bdevs_list": [ 00:09:44.990 { 00:09:44.990 "name": null, 00:09:44.990 "uuid": "9131a959-6e0a-4dc6-a386-ed259b3a7a20", 00:09:44.990 "is_configured": false, 00:09:44.990 "data_offset": 0, 00:09:44.990 "data_size": 63488 00:09:44.990 }, 00:09:44.990 { 00:09:44.990 "name": null, 00:09:44.990 "uuid": "f1bafcc1-228d-498c-8b2e-a9af417030e4", 00:09:44.990 "is_configured": false, 00:09:44.990 "data_offset": 0, 00:09:44.990 "data_size": 63488 00:09:44.990 }, 00:09:44.990 { 00:09:44.990 "name": "BaseBdev3", 00:09:44.990 "uuid": "cf93741b-f85f-4200-b240-a2d3937d1479", 00:09:44.990 "is_configured": true, 00:09:44.990 "data_offset": 2048, 00:09:44.990 "data_size": 63488 00:09:44.990 } 00:09:44.990 ] 00:09:44.990 }' 00:09:44.990 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:44.990 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:45.247 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:45.247 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.247 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.247 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:45.247 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:45.505 [2024-11-20 15:57:43.507294] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:45.505 "name": "Existed_Raid", 00:09:45.505 "uuid": "2db3b179-1bf8-4b92-b22c-493cbaf50220", 00:09:45.505 "strip_size_kb": 64, 00:09:45.505 "state": "configuring", 00:09:45.505 "raid_level": "raid0", 00:09:45.505 "superblock": true, 00:09:45.505 "num_base_bdevs": 3, 00:09:45.505 "num_base_bdevs_discovered": 2, 00:09:45.505 "num_base_bdevs_operational": 3, 00:09:45.505 "base_bdevs_list": [ 00:09:45.505 { 00:09:45.505 "name": null, 00:09:45.505 "uuid": "9131a959-6e0a-4dc6-a386-ed259b3a7a20", 00:09:45.505 "is_configured": false, 00:09:45.505 "data_offset": 0, 00:09:45.505 "data_size": 63488 00:09:45.505 }, 00:09:45.505 { 00:09:45.505 "name": "BaseBdev2", 00:09:45.505 "uuid": "f1bafcc1-228d-498c-8b2e-a9af417030e4", 00:09:45.505 "is_configured": true, 00:09:45.505 "data_offset": 2048, 00:09:45.505 "data_size": 63488 00:09:45.505 }, 00:09:45.505 { 00:09:45.505 "name": "BaseBdev3", 00:09:45.505 "uuid": "cf93741b-f85f-4200-b240-a2d3937d1479", 00:09:45.505 "is_configured": true, 00:09:45.505 "data_offset": 2048, 00:09:45.505 "data_size": 63488 00:09:45.505 } 00:09:45.505 ] 00:09:45.505 }' 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:45.505 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 9131a959-6e0a-4dc6-a386-ed259b3a7a20 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:45.763 [2024-11-20 15:57:43.905535] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:45.763 NewBaseBdev 00:09:45.763 [2024-11-20 15:57:43.905798] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:09:45.763 [2024-11-20 15:57:43.905816] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:45.763 [2024-11-20 15:57:43.906016] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:09:45.763 [2024-11-20 15:57:43.906113] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:09:45.763 [2024-11-20 15:57:43.906119] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.763 [2024-11-20 15:57:43.906211] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:45.763 [ 00:09:45.763 { 00:09:45.763 "name": "NewBaseBdev", 00:09:45.763 "aliases": [ 00:09:45.763 "9131a959-6e0a-4dc6-a386-ed259b3a7a20" 00:09:45.763 ], 00:09:45.763 "product_name": "Malloc disk", 00:09:45.763 "block_size": 512, 00:09:45.763 "num_blocks": 65536, 00:09:45.763 "uuid": "9131a959-6e0a-4dc6-a386-ed259b3a7a20", 00:09:45.763 "assigned_rate_limits": { 00:09:45.763 "rw_ios_per_sec": 0, 00:09:45.763 "rw_mbytes_per_sec": 0, 00:09:45.763 "r_mbytes_per_sec": 0, 00:09:45.763 "w_mbytes_per_sec": 0 00:09:45.763 }, 00:09:45.763 "claimed": true, 00:09:45.763 "claim_type": "exclusive_write", 00:09:45.763 "zoned": false, 00:09:45.763 "supported_io_types": { 00:09:45.763 "read": true, 00:09:45.763 "write": true, 00:09:45.763 "unmap": true, 00:09:45.763 "flush": true, 00:09:45.763 "reset": true, 00:09:45.763 "nvme_admin": false, 00:09:45.763 "nvme_io": false, 00:09:45.763 "nvme_io_md": false, 00:09:45.763 "write_zeroes": true, 00:09:45.763 "zcopy": true, 00:09:45.763 "get_zone_info": false, 00:09:45.763 "zone_management": false, 00:09:45.763 "zone_append": false, 00:09:45.763 "compare": false, 00:09:45.763 "compare_and_write": false, 00:09:45.763 "abort": true, 00:09:45.763 "seek_hole": false, 00:09:45.763 "seek_data": false, 00:09:45.763 "copy": true, 00:09:45.763 "nvme_iov_md": false 00:09:45.763 }, 00:09:45.763 "memory_domains": [ 00:09:45.763 { 00:09:45.763 "dma_device_id": "system", 00:09:45.763 "dma_device_type": 1 00:09:45.763 }, 00:09:45.763 { 00:09:45.763 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:45.763 "dma_device_type": 2 00:09:45.763 } 00:09:45.763 ], 00:09:45.763 "driver_specific": {} 00:09:45.763 } 00:09:45.763 ] 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:45.763 "name": "Existed_Raid", 00:09:45.763 "uuid": "2db3b179-1bf8-4b92-b22c-493cbaf50220", 00:09:45.763 "strip_size_kb": 64, 00:09:45.763 "state": "online", 00:09:45.763 "raid_level": "raid0", 00:09:45.763 "superblock": true, 00:09:45.763 "num_base_bdevs": 3, 00:09:45.763 "num_base_bdevs_discovered": 3, 00:09:45.763 "num_base_bdevs_operational": 3, 00:09:45.763 "base_bdevs_list": [ 00:09:45.763 { 00:09:45.763 "name": "NewBaseBdev", 00:09:45.763 "uuid": "9131a959-6e0a-4dc6-a386-ed259b3a7a20", 00:09:45.763 "is_configured": true, 00:09:45.763 "data_offset": 2048, 00:09:45.763 "data_size": 63488 00:09:45.763 }, 00:09:45.763 { 00:09:45.763 "name": "BaseBdev2", 00:09:45.763 "uuid": "f1bafcc1-228d-498c-8b2e-a9af417030e4", 00:09:45.763 "is_configured": true, 00:09:45.763 "data_offset": 2048, 00:09:45.763 "data_size": 63488 00:09:45.763 }, 00:09:45.763 { 00:09:45.763 "name": "BaseBdev3", 00:09:45.763 "uuid": "cf93741b-f85f-4200-b240-a2d3937d1479", 00:09:45.763 "is_configured": true, 00:09:45.763 "data_offset": 2048, 00:09:45.763 "data_size": 63488 00:09:45.763 } 00:09:45.763 ] 00:09:45.763 }' 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:45.763 15:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:46.021 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:46.021 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:46.021 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:46.021 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:46.021 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:46.021 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:46.021 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:46.021 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:46.021 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.021 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:46.021 [2024-11-20 15:57:44.209897] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:46.021 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.021 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:46.021 "name": "Existed_Raid", 00:09:46.021 "aliases": [ 00:09:46.021 "2db3b179-1bf8-4b92-b22c-493cbaf50220" 00:09:46.021 ], 00:09:46.021 "product_name": "Raid Volume", 00:09:46.021 "block_size": 512, 00:09:46.021 "num_blocks": 190464, 00:09:46.021 "uuid": "2db3b179-1bf8-4b92-b22c-493cbaf50220", 00:09:46.021 "assigned_rate_limits": { 00:09:46.021 "rw_ios_per_sec": 0, 00:09:46.021 "rw_mbytes_per_sec": 0, 00:09:46.021 "r_mbytes_per_sec": 0, 00:09:46.021 "w_mbytes_per_sec": 0 00:09:46.021 }, 00:09:46.021 "claimed": false, 00:09:46.021 "zoned": false, 00:09:46.021 "supported_io_types": { 00:09:46.021 "read": true, 00:09:46.021 "write": true, 00:09:46.021 "unmap": true, 00:09:46.021 "flush": true, 00:09:46.021 "reset": true, 00:09:46.021 "nvme_admin": false, 00:09:46.021 "nvme_io": false, 00:09:46.021 "nvme_io_md": false, 00:09:46.021 "write_zeroes": true, 00:09:46.021 "zcopy": false, 00:09:46.021 "get_zone_info": false, 00:09:46.021 "zone_management": false, 00:09:46.021 "zone_append": false, 00:09:46.021 "compare": false, 00:09:46.021 "compare_and_write": false, 00:09:46.021 "abort": false, 00:09:46.021 "seek_hole": false, 00:09:46.021 "seek_data": false, 00:09:46.021 "copy": false, 00:09:46.021 "nvme_iov_md": false 00:09:46.021 }, 00:09:46.021 "memory_domains": [ 00:09:46.021 { 00:09:46.021 "dma_device_id": "system", 00:09:46.021 "dma_device_type": 1 00:09:46.021 }, 00:09:46.021 { 00:09:46.021 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:46.021 "dma_device_type": 2 00:09:46.021 }, 00:09:46.021 { 00:09:46.021 "dma_device_id": "system", 00:09:46.021 "dma_device_type": 1 00:09:46.021 }, 00:09:46.021 { 00:09:46.021 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:46.021 "dma_device_type": 2 00:09:46.021 }, 00:09:46.021 { 00:09:46.021 "dma_device_id": "system", 00:09:46.021 "dma_device_type": 1 00:09:46.021 }, 00:09:46.021 { 00:09:46.021 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:46.021 "dma_device_type": 2 00:09:46.021 } 00:09:46.021 ], 00:09:46.021 "driver_specific": { 00:09:46.021 "raid": { 00:09:46.021 "uuid": "2db3b179-1bf8-4b92-b22c-493cbaf50220", 00:09:46.021 "strip_size_kb": 64, 00:09:46.021 "state": "online", 00:09:46.021 "raid_level": "raid0", 00:09:46.021 "superblock": true, 00:09:46.021 "num_base_bdevs": 3, 00:09:46.021 "num_base_bdevs_discovered": 3, 00:09:46.021 "num_base_bdevs_operational": 3, 00:09:46.021 "base_bdevs_list": [ 00:09:46.021 { 00:09:46.021 "name": "NewBaseBdev", 00:09:46.021 "uuid": "9131a959-6e0a-4dc6-a386-ed259b3a7a20", 00:09:46.021 "is_configured": true, 00:09:46.021 "data_offset": 2048, 00:09:46.021 "data_size": 63488 00:09:46.021 }, 00:09:46.021 { 00:09:46.021 "name": "BaseBdev2", 00:09:46.021 "uuid": "f1bafcc1-228d-498c-8b2e-a9af417030e4", 00:09:46.021 "is_configured": true, 00:09:46.021 "data_offset": 2048, 00:09:46.021 "data_size": 63488 00:09:46.021 }, 00:09:46.021 { 00:09:46.021 "name": "BaseBdev3", 00:09:46.022 "uuid": "cf93741b-f85f-4200-b240-a2d3937d1479", 00:09:46.022 "is_configured": true, 00:09:46.022 "data_offset": 2048, 00:09:46.022 "data_size": 63488 00:09:46.022 } 00:09:46.022 ] 00:09:46.022 } 00:09:46.022 } 00:09:46.022 }' 00:09:46.022 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:46.280 BaseBdev2 00:09:46.280 BaseBdev3' 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:46.280 [2024-11-20 15:57:44.397660] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:46.280 [2024-11-20 15:57:44.397688] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:46.280 [2024-11-20 15:57:44.397740] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:46.280 [2024-11-20 15:57:44.397787] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:46.280 [2024-11-20 15:57:44.397796] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 62922 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 62922 ']' 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 62922 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:46.280 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 62922 00:09:46.281 killing process with pid 62922 00:09:46.281 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:46.281 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:46.281 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 62922' 00:09:46.281 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 62922 00:09:46.281 [2024-11-20 15:57:44.424543] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:46.281 15:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 62922 00:09:46.538 [2024-11-20 15:57:44.572316] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:47.106 15:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:47.106 00:09:47.106 real 0m7.398s 00:09:47.106 user 0m11.975s 00:09:47.106 sys 0m1.166s 00:09:47.106 15:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:47.106 ************************************ 00:09:47.106 END TEST raid_state_function_test_sb 00:09:47.106 ************************************ 00:09:47.106 15:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:47.106 15:57:45 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 3 00:09:47.106 15:57:45 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:09:47.106 15:57:45 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:47.106 15:57:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:47.106 ************************************ 00:09:47.106 START TEST raid_superblock_test 00:09:47.106 ************************************ 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 3 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:09:47.106 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=63509 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 63509 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 63509 ']' 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.106 15:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:47.106 [2024-11-20 15:57:45.261181] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:09:47.106 [2024-11-20 15:57:45.261464] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid63509 ] 00:09:47.367 [2024-11-20 15:57:45.419293] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:47.367 [2024-11-20 15:57:45.520179] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:47.626 [2024-11-20 15:57:45.656475] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:47.626 [2024-11-20 15:57:45.656659] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:47.895 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:47.895 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:09:47.895 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:47.895 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:47.895 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:47.895 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:47.895 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:47.895 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:47.895 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:47.895 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:47.895 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:47.895 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.895 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.155 malloc1 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.155 [2024-11-20 15:57:46.166730] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:48.155 [2024-11-20 15:57:46.166787] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:48.155 [2024-11-20 15:57:46.166808] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:48.155 [2024-11-20 15:57:46.166818] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:48.155 [2024-11-20 15:57:46.168952] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:48.155 [2024-11-20 15:57:46.169085] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:48.155 pt1 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.155 malloc2 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.155 [2024-11-20 15:57:46.202529] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:48.155 [2024-11-20 15:57:46.202592] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:48.155 [2024-11-20 15:57:46.202617] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:48.155 [2024-11-20 15:57:46.202626] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:48.155 [2024-11-20 15:57:46.204720] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:48.155 [2024-11-20 15:57:46.204750] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:48.155 pt2 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.155 malloc3 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.155 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.155 [2024-11-20 15:57:46.253794] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:48.155 [2024-11-20 15:57:46.253845] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:48.155 [2024-11-20 15:57:46.253866] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:48.155 [2024-11-20 15:57:46.253875] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:48.156 [2024-11-20 15:57:46.255965] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:48.156 [2024-11-20 15:57:46.255996] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:48.156 pt3 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.156 [2024-11-20 15:57:46.265842] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:48.156 [2024-11-20 15:57:46.267658] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:48.156 [2024-11-20 15:57:46.267733] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:48.156 [2024-11-20 15:57:46.267882] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:09:48.156 [2024-11-20 15:57:46.267894] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:48.156 [2024-11-20 15:57:46.268134] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:09:48.156 [2024-11-20 15:57:46.268270] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:09:48.156 [2024-11-20 15:57:46.268278] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:09:48.156 [2024-11-20 15:57:46.268409] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:48.156 "name": "raid_bdev1", 00:09:48.156 "uuid": "58713545-af5d-406e-82e7-64698e673c19", 00:09:48.156 "strip_size_kb": 64, 00:09:48.156 "state": "online", 00:09:48.156 "raid_level": "raid0", 00:09:48.156 "superblock": true, 00:09:48.156 "num_base_bdevs": 3, 00:09:48.156 "num_base_bdevs_discovered": 3, 00:09:48.156 "num_base_bdevs_operational": 3, 00:09:48.156 "base_bdevs_list": [ 00:09:48.156 { 00:09:48.156 "name": "pt1", 00:09:48.156 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:48.156 "is_configured": true, 00:09:48.156 "data_offset": 2048, 00:09:48.156 "data_size": 63488 00:09:48.156 }, 00:09:48.156 { 00:09:48.156 "name": "pt2", 00:09:48.156 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:48.156 "is_configured": true, 00:09:48.156 "data_offset": 2048, 00:09:48.156 "data_size": 63488 00:09:48.156 }, 00:09:48.156 { 00:09:48.156 "name": "pt3", 00:09:48.156 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:48.156 "is_configured": true, 00:09:48.156 "data_offset": 2048, 00:09:48.156 "data_size": 63488 00:09:48.156 } 00:09:48.156 ] 00:09:48.156 }' 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:48.156 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.414 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:48.414 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:48.415 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:48.415 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:48.415 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:48.415 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:48.415 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:48.415 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.415 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.415 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:48.415 [2024-11-20 15:57:46.610221] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:48.415 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.415 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:48.415 "name": "raid_bdev1", 00:09:48.415 "aliases": [ 00:09:48.415 "58713545-af5d-406e-82e7-64698e673c19" 00:09:48.415 ], 00:09:48.415 "product_name": "Raid Volume", 00:09:48.415 "block_size": 512, 00:09:48.415 "num_blocks": 190464, 00:09:48.415 "uuid": "58713545-af5d-406e-82e7-64698e673c19", 00:09:48.415 "assigned_rate_limits": { 00:09:48.415 "rw_ios_per_sec": 0, 00:09:48.415 "rw_mbytes_per_sec": 0, 00:09:48.415 "r_mbytes_per_sec": 0, 00:09:48.415 "w_mbytes_per_sec": 0 00:09:48.415 }, 00:09:48.415 "claimed": false, 00:09:48.415 "zoned": false, 00:09:48.415 "supported_io_types": { 00:09:48.415 "read": true, 00:09:48.415 "write": true, 00:09:48.415 "unmap": true, 00:09:48.415 "flush": true, 00:09:48.415 "reset": true, 00:09:48.415 "nvme_admin": false, 00:09:48.415 "nvme_io": false, 00:09:48.415 "nvme_io_md": false, 00:09:48.415 "write_zeroes": true, 00:09:48.415 "zcopy": false, 00:09:48.415 "get_zone_info": false, 00:09:48.415 "zone_management": false, 00:09:48.415 "zone_append": false, 00:09:48.415 "compare": false, 00:09:48.415 "compare_and_write": false, 00:09:48.415 "abort": false, 00:09:48.415 "seek_hole": false, 00:09:48.415 "seek_data": false, 00:09:48.415 "copy": false, 00:09:48.415 "nvme_iov_md": false 00:09:48.415 }, 00:09:48.415 "memory_domains": [ 00:09:48.415 { 00:09:48.415 "dma_device_id": "system", 00:09:48.415 "dma_device_type": 1 00:09:48.415 }, 00:09:48.415 { 00:09:48.415 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:48.415 "dma_device_type": 2 00:09:48.415 }, 00:09:48.415 { 00:09:48.415 "dma_device_id": "system", 00:09:48.415 "dma_device_type": 1 00:09:48.415 }, 00:09:48.415 { 00:09:48.415 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:48.415 "dma_device_type": 2 00:09:48.415 }, 00:09:48.415 { 00:09:48.415 "dma_device_id": "system", 00:09:48.415 "dma_device_type": 1 00:09:48.415 }, 00:09:48.415 { 00:09:48.415 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:48.415 "dma_device_type": 2 00:09:48.415 } 00:09:48.415 ], 00:09:48.415 "driver_specific": { 00:09:48.415 "raid": { 00:09:48.415 "uuid": "58713545-af5d-406e-82e7-64698e673c19", 00:09:48.415 "strip_size_kb": 64, 00:09:48.415 "state": "online", 00:09:48.415 "raid_level": "raid0", 00:09:48.415 "superblock": true, 00:09:48.415 "num_base_bdevs": 3, 00:09:48.415 "num_base_bdevs_discovered": 3, 00:09:48.415 "num_base_bdevs_operational": 3, 00:09:48.415 "base_bdevs_list": [ 00:09:48.415 { 00:09:48.415 "name": "pt1", 00:09:48.415 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:48.415 "is_configured": true, 00:09:48.415 "data_offset": 2048, 00:09:48.415 "data_size": 63488 00:09:48.415 }, 00:09:48.415 { 00:09:48.415 "name": "pt2", 00:09:48.415 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:48.415 "is_configured": true, 00:09:48.415 "data_offset": 2048, 00:09:48.415 "data_size": 63488 00:09:48.415 }, 00:09:48.415 { 00:09:48.415 "name": "pt3", 00:09:48.415 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:48.415 "is_configured": true, 00:09:48.415 "data_offset": 2048, 00:09:48.415 "data_size": 63488 00:09:48.415 } 00:09:48.415 ] 00:09:48.415 } 00:09:48.415 } 00:09:48.415 }' 00:09:48.415 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:48.677 pt2 00:09:48.677 pt3' 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:48.677 [2024-11-20 15:57:46.810227] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=58713545-af5d-406e-82e7-64698e673c19 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 58713545-af5d-406e-82e7-64698e673c19 ']' 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.677 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.678 [2024-11-20 15:57:46.833927] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:48.678 [2024-11-20 15:57:46.833955] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:48.678 [2024-11-20 15:57:46.834019] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:48.678 [2024-11-20 15:57:46.834082] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:48.678 [2024-11-20 15:57:46.834092] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.678 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.938 [2024-11-20 15:57:46.934002] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:48.938 [2024-11-20 15:57:46.935874] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:48.938 [2024-11-20 15:57:46.935920] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:48.938 [2024-11-20 15:57:46.935967] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:48.938 [2024-11-20 15:57:46.936012] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:48.938 [2024-11-20 15:57:46.936031] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:48.938 [2024-11-20 15:57:46.936047] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:48.938 [2024-11-20 15:57:46.936058] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:09:48.938 request: 00:09:48.938 { 00:09:48.938 "name": "raid_bdev1", 00:09:48.938 "raid_level": "raid0", 00:09:48.938 "base_bdevs": [ 00:09:48.938 "malloc1", 00:09:48.938 "malloc2", 00:09:48.938 "malloc3" 00:09:48.938 ], 00:09:48.938 "strip_size_kb": 64, 00:09:48.938 "superblock": false, 00:09:48.938 "method": "bdev_raid_create", 00:09:48.938 "req_id": 1 00:09:48.938 } 00:09:48.938 Got JSON-RPC error response 00:09:48.938 response: 00:09:48.938 { 00:09:48.938 "code": -17, 00:09:48.938 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:48.938 } 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.938 [2024-11-20 15:57:46.981967] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:48.938 [2024-11-20 15:57:46.982088] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:48.938 [2024-11-20 15:57:46.982126] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:48.938 [2024-11-20 15:57:46.982185] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:48.938 [2024-11-20 15:57:46.984375] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:48.938 [2024-11-20 15:57:46.984481] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:48.938 [2024-11-20 15:57:46.984562] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:48.938 [2024-11-20 15:57:46.984611] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:48.938 pt1 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:48.938 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:48.939 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:48.939 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:48.939 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:48.939 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:48.939 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:48.939 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:48.939 15:57:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.939 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.939 15:57:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.939 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.939 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:48.939 "name": "raid_bdev1", 00:09:48.939 "uuid": "58713545-af5d-406e-82e7-64698e673c19", 00:09:48.939 "strip_size_kb": 64, 00:09:48.939 "state": "configuring", 00:09:48.939 "raid_level": "raid0", 00:09:48.939 "superblock": true, 00:09:48.939 "num_base_bdevs": 3, 00:09:48.939 "num_base_bdevs_discovered": 1, 00:09:48.939 "num_base_bdevs_operational": 3, 00:09:48.939 "base_bdevs_list": [ 00:09:48.939 { 00:09:48.939 "name": "pt1", 00:09:48.939 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:48.939 "is_configured": true, 00:09:48.939 "data_offset": 2048, 00:09:48.939 "data_size": 63488 00:09:48.939 }, 00:09:48.939 { 00:09:48.939 "name": null, 00:09:48.939 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:48.939 "is_configured": false, 00:09:48.939 "data_offset": 2048, 00:09:48.939 "data_size": 63488 00:09:48.939 }, 00:09:48.939 { 00:09:48.939 "name": null, 00:09:48.939 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:48.939 "is_configured": false, 00:09:48.939 "data_offset": 2048, 00:09:48.939 "data_size": 63488 00:09:48.939 } 00:09:48.939 ] 00:09:48.939 }' 00:09:48.939 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:48.939 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.199 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:09:49.199 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:49.199 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.199 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.199 [2024-11-20 15:57:47.314087] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:49.199 [2024-11-20 15:57:47.314144] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:49.199 [2024-11-20 15:57:47.314167] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:09:49.199 [2024-11-20 15:57:47.314177] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:49.199 [2024-11-20 15:57:47.314569] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:49.199 [2024-11-20 15:57:47.314582] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:49.199 [2024-11-20 15:57:47.314653] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:49.199 [2024-11-20 15:57:47.314688] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:49.199 pt2 00:09:49.199 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.199 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:09:49.199 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.199 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.199 [2024-11-20 15:57:47.322091] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:09:49.199 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.199 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:09:49.199 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:49.199 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:49.199 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:49.199 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:49.199 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:49.200 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:49.200 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:49.200 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:49.200 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:49.200 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.200 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.200 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.200 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:49.200 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.200 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:49.200 "name": "raid_bdev1", 00:09:49.200 "uuid": "58713545-af5d-406e-82e7-64698e673c19", 00:09:49.200 "strip_size_kb": 64, 00:09:49.200 "state": "configuring", 00:09:49.200 "raid_level": "raid0", 00:09:49.200 "superblock": true, 00:09:49.200 "num_base_bdevs": 3, 00:09:49.200 "num_base_bdevs_discovered": 1, 00:09:49.200 "num_base_bdevs_operational": 3, 00:09:49.200 "base_bdevs_list": [ 00:09:49.200 { 00:09:49.200 "name": "pt1", 00:09:49.200 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:49.200 "is_configured": true, 00:09:49.200 "data_offset": 2048, 00:09:49.200 "data_size": 63488 00:09:49.200 }, 00:09:49.200 { 00:09:49.200 "name": null, 00:09:49.200 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:49.200 "is_configured": false, 00:09:49.200 "data_offset": 0, 00:09:49.200 "data_size": 63488 00:09:49.200 }, 00:09:49.200 { 00:09:49.200 "name": null, 00:09:49.200 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:49.200 "is_configured": false, 00:09:49.200 "data_offset": 2048, 00:09:49.200 "data_size": 63488 00:09:49.200 } 00:09:49.200 ] 00:09:49.200 }' 00:09:49.200 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:49.200 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.457 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:49.457 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:49.457 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:49.457 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.457 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.457 [2024-11-20 15:57:47.634144] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:49.457 [2024-11-20 15:57:47.634199] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:49.458 [2024-11-20 15:57:47.634214] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:09:49.458 [2024-11-20 15:57:47.634224] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:49.458 [2024-11-20 15:57:47.634622] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:49.458 [2024-11-20 15:57:47.634637] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:49.458 [2024-11-20 15:57:47.634715] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:49.458 [2024-11-20 15:57:47.634736] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:49.458 pt2 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.458 [2024-11-20 15:57:47.642137] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:49.458 [2024-11-20 15:57:47.642177] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:49.458 [2024-11-20 15:57:47.642188] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:49.458 [2024-11-20 15:57:47.642198] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:49.458 [2024-11-20 15:57:47.642538] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:49.458 [2024-11-20 15:57:47.642554] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:49.458 [2024-11-20 15:57:47.642606] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:49.458 [2024-11-20 15:57:47.642623] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:49.458 [2024-11-20 15:57:47.642743] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:09:49.458 [2024-11-20 15:57:47.642755] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:49.458 [2024-11-20 15:57:47.642984] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:09:49.458 [2024-11-20 15:57:47.643111] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:09:49.458 [2024-11-20 15:57:47.643119] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:09:49.458 [2024-11-20 15:57:47.643239] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:49.458 pt3 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:49.458 "name": "raid_bdev1", 00:09:49.458 "uuid": "58713545-af5d-406e-82e7-64698e673c19", 00:09:49.458 "strip_size_kb": 64, 00:09:49.458 "state": "online", 00:09:49.458 "raid_level": "raid0", 00:09:49.458 "superblock": true, 00:09:49.458 "num_base_bdevs": 3, 00:09:49.458 "num_base_bdevs_discovered": 3, 00:09:49.458 "num_base_bdevs_operational": 3, 00:09:49.458 "base_bdevs_list": [ 00:09:49.458 { 00:09:49.458 "name": "pt1", 00:09:49.458 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:49.458 "is_configured": true, 00:09:49.458 "data_offset": 2048, 00:09:49.458 "data_size": 63488 00:09:49.458 }, 00:09:49.458 { 00:09:49.458 "name": "pt2", 00:09:49.458 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:49.458 "is_configured": true, 00:09:49.458 "data_offset": 2048, 00:09:49.458 "data_size": 63488 00:09:49.458 }, 00:09:49.458 { 00:09:49.458 "name": "pt3", 00:09:49.458 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:49.458 "is_configured": true, 00:09:49.458 "data_offset": 2048, 00:09:49.458 "data_size": 63488 00:09:49.458 } 00:09:49.458 ] 00:09:49.458 }' 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:49.458 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.715 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:49.715 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:49.715 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:49.715 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:49.715 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:49.715 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:49.715 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:49.715 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.715 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.715 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:49.715 [2024-11-20 15:57:47.950545] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:49.715 15:57:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.976 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:49.976 "name": "raid_bdev1", 00:09:49.976 "aliases": [ 00:09:49.976 "58713545-af5d-406e-82e7-64698e673c19" 00:09:49.976 ], 00:09:49.976 "product_name": "Raid Volume", 00:09:49.976 "block_size": 512, 00:09:49.976 "num_blocks": 190464, 00:09:49.976 "uuid": "58713545-af5d-406e-82e7-64698e673c19", 00:09:49.976 "assigned_rate_limits": { 00:09:49.976 "rw_ios_per_sec": 0, 00:09:49.976 "rw_mbytes_per_sec": 0, 00:09:49.976 "r_mbytes_per_sec": 0, 00:09:49.976 "w_mbytes_per_sec": 0 00:09:49.976 }, 00:09:49.976 "claimed": false, 00:09:49.976 "zoned": false, 00:09:49.976 "supported_io_types": { 00:09:49.976 "read": true, 00:09:49.976 "write": true, 00:09:49.976 "unmap": true, 00:09:49.976 "flush": true, 00:09:49.976 "reset": true, 00:09:49.976 "nvme_admin": false, 00:09:49.976 "nvme_io": false, 00:09:49.976 "nvme_io_md": false, 00:09:49.976 "write_zeroes": true, 00:09:49.976 "zcopy": false, 00:09:49.976 "get_zone_info": false, 00:09:49.976 "zone_management": false, 00:09:49.976 "zone_append": false, 00:09:49.976 "compare": false, 00:09:49.976 "compare_and_write": false, 00:09:49.976 "abort": false, 00:09:49.976 "seek_hole": false, 00:09:49.976 "seek_data": false, 00:09:49.976 "copy": false, 00:09:49.976 "nvme_iov_md": false 00:09:49.976 }, 00:09:49.976 "memory_domains": [ 00:09:49.976 { 00:09:49.976 "dma_device_id": "system", 00:09:49.976 "dma_device_type": 1 00:09:49.976 }, 00:09:49.976 { 00:09:49.976 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:49.976 "dma_device_type": 2 00:09:49.976 }, 00:09:49.976 { 00:09:49.976 "dma_device_id": "system", 00:09:49.976 "dma_device_type": 1 00:09:49.976 }, 00:09:49.976 { 00:09:49.976 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:49.976 "dma_device_type": 2 00:09:49.976 }, 00:09:49.976 { 00:09:49.976 "dma_device_id": "system", 00:09:49.976 "dma_device_type": 1 00:09:49.976 }, 00:09:49.976 { 00:09:49.976 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:49.976 "dma_device_type": 2 00:09:49.976 } 00:09:49.976 ], 00:09:49.976 "driver_specific": { 00:09:49.976 "raid": { 00:09:49.976 "uuid": "58713545-af5d-406e-82e7-64698e673c19", 00:09:49.976 "strip_size_kb": 64, 00:09:49.976 "state": "online", 00:09:49.976 "raid_level": "raid0", 00:09:49.976 "superblock": true, 00:09:49.976 "num_base_bdevs": 3, 00:09:49.976 "num_base_bdevs_discovered": 3, 00:09:49.976 "num_base_bdevs_operational": 3, 00:09:49.976 "base_bdevs_list": [ 00:09:49.976 { 00:09:49.976 "name": "pt1", 00:09:49.976 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:49.976 "is_configured": true, 00:09:49.976 "data_offset": 2048, 00:09:49.976 "data_size": 63488 00:09:49.976 }, 00:09:49.976 { 00:09:49.976 "name": "pt2", 00:09:49.976 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:49.976 "is_configured": true, 00:09:49.976 "data_offset": 2048, 00:09:49.976 "data_size": 63488 00:09:49.976 }, 00:09:49.976 { 00:09:49.976 "name": "pt3", 00:09:49.976 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:49.976 "is_configured": true, 00:09:49.976 "data_offset": 2048, 00:09:49.976 "data_size": 63488 00:09:49.976 } 00:09:49.976 ] 00:09:49.976 } 00:09:49.976 } 00:09:49.976 }' 00:09:49.976 15:57:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:49.976 pt2 00:09:49.976 pt3' 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:49.976 [2024-11-20 15:57:48.154558] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.976 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 58713545-af5d-406e-82e7-64698e673c19 '!=' 58713545-af5d-406e-82e7-64698e673c19 ']' 00:09:49.977 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:09:49.977 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:49.977 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:49.977 15:57:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 63509 00:09:49.977 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 63509 ']' 00:09:49.977 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 63509 00:09:49.977 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:09:49.977 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:49.977 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 63509 00:09:49.977 killing process with pid 63509 00:09:49.977 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:49.977 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:49.977 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 63509' 00:09:49.977 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 63509 00:09:49.977 [2024-11-20 15:57:48.203956] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:49.977 15:57:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 63509 00:09:49.977 [2024-11-20 15:57:48.204042] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:49.977 [2024-11-20 15:57:48.204103] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:49.977 [2024-11-20 15:57:48.204115] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:09:50.236 [2024-11-20 15:57:48.391932] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:51.178 15:57:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:51.178 00:09:51.178 real 0m3.901s 00:09:51.178 user 0m5.650s 00:09:51.178 sys 0m0.557s 00:09:51.178 ************************************ 00:09:51.178 END TEST raid_superblock_test 00:09:51.178 ************************************ 00:09:51.178 15:57:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:51.178 15:57:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.178 15:57:49 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 3 read 00:09:51.178 15:57:49 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:51.178 15:57:49 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:51.178 15:57:49 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:51.178 ************************************ 00:09:51.178 START TEST raid_read_error_test 00:09:51.178 ************************************ 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 3 read 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:51.178 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.QaujsBFrft 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=63751 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 63751 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 63751 ']' 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.178 15:57:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:51.178 [2024-11-20 15:57:49.210632] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:09:51.178 [2024-11-20 15:57:49.210775] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid63751 ] 00:09:51.457 [2024-11-20 15:57:49.364792] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:51.457 [2024-11-20 15:57:49.463483] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:51.457 [2024-11-20 15:57:49.600206] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:51.457 [2024-11-20 15:57:49.600382] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.037 BaseBdev1_malloc 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.037 true 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.037 [2024-11-20 15:57:50.097652] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:52.037 [2024-11-20 15:57:50.097715] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:52.037 [2024-11-20 15:57:50.097733] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:52.037 [2024-11-20 15:57:50.097745] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:52.037 [2024-11-20 15:57:50.099850] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:52.037 [2024-11-20 15:57:50.099888] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:52.037 BaseBdev1 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.037 BaseBdev2_malloc 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.037 true 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.037 [2024-11-20 15:57:50.141450] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:52.037 [2024-11-20 15:57:50.141494] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:52.037 [2024-11-20 15:57:50.141509] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:52.037 [2024-11-20 15:57:50.141519] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:52.037 [2024-11-20 15:57:50.143588] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:52.037 [2024-11-20 15:57:50.143746] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:52.037 BaseBdev2 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.037 BaseBdev3_malloc 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.037 true 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.037 [2024-11-20 15:57:50.196393] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:52.037 [2024-11-20 15:57:50.196443] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:52.037 [2024-11-20 15:57:50.196460] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:52.037 [2024-11-20 15:57:50.196471] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:52.037 [2024-11-20 15:57:50.198583] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:52.037 [2024-11-20 15:57:50.198619] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:52.037 BaseBdev3 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.037 [2024-11-20 15:57:50.204467] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:52.037 [2024-11-20 15:57:50.206432] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:52.037 [2024-11-20 15:57:50.206588] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:52.037 [2024-11-20 15:57:50.206914] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:09:52.037 [2024-11-20 15:57:50.206994] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:52.037 [2024-11-20 15:57:50.207260] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:09:52.037 [2024-11-20 15:57:50.207474] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:09:52.037 [2024-11-20 15:57:50.207506] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:09:52.037 [2024-11-20 15:57:50.207722] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:09:52.037 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:52.038 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:52.038 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:52.038 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:52.038 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:52.038 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:52.038 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:52.038 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:52.038 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:52.038 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.038 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:52.038 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.038 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.038 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.038 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:52.038 "name": "raid_bdev1", 00:09:52.038 "uuid": "44771f2b-2178-4553-b5a9-e867f37d9956", 00:09:52.038 "strip_size_kb": 64, 00:09:52.038 "state": "online", 00:09:52.038 "raid_level": "raid0", 00:09:52.038 "superblock": true, 00:09:52.038 "num_base_bdevs": 3, 00:09:52.038 "num_base_bdevs_discovered": 3, 00:09:52.038 "num_base_bdevs_operational": 3, 00:09:52.038 "base_bdevs_list": [ 00:09:52.038 { 00:09:52.038 "name": "BaseBdev1", 00:09:52.038 "uuid": "385d8332-a10c-5f3f-a813-a249e80e01ad", 00:09:52.038 "is_configured": true, 00:09:52.038 "data_offset": 2048, 00:09:52.038 "data_size": 63488 00:09:52.038 }, 00:09:52.038 { 00:09:52.038 "name": "BaseBdev2", 00:09:52.038 "uuid": "962dff4f-e94b-5273-ae06-aa332bef910d", 00:09:52.038 "is_configured": true, 00:09:52.038 "data_offset": 2048, 00:09:52.038 "data_size": 63488 00:09:52.038 }, 00:09:52.038 { 00:09:52.038 "name": "BaseBdev3", 00:09:52.038 "uuid": "9de0a78b-d84f-5612-bdf1-9be7487e0e2b", 00:09:52.038 "is_configured": true, 00:09:52.038 "data_offset": 2048, 00:09:52.038 "data_size": 63488 00:09:52.038 } 00:09:52.038 ] 00:09:52.038 }' 00:09:52.038 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:52.038 15:57:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.294 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:52.294 15:57:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:52.550 [2024-11-20 15:57:50.601491] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:53.480 "name": "raid_bdev1", 00:09:53.480 "uuid": "44771f2b-2178-4553-b5a9-e867f37d9956", 00:09:53.480 "strip_size_kb": 64, 00:09:53.480 "state": "online", 00:09:53.480 "raid_level": "raid0", 00:09:53.480 "superblock": true, 00:09:53.480 "num_base_bdevs": 3, 00:09:53.480 "num_base_bdevs_discovered": 3, 00:09:53.480 "num_base_bdevs_operational": 3, 00:09:53.480 "base_bdevs_list": [ 00:09:53.480 { 00:09:53.480 "name": "BaseBdev1", 00:09:53.480 "uuid": "385d8332-a10c-5f3f-a813-a249e80e01ad", 00:09:53.480 "is_configured": true, 00:09:53.480 "data_offset": 2048, 00:09:53.480 "data_size": 63488 00:09:53.480 }, 00:09:53.480 { 00:09:53.480 "name": "BaseBdev2", 00:09:53.480 "uuid": "962dff4f-e94b-5273-ae06-aa332bef910d", 00:09:53.480 "is_configured": true, 00:09:53.480 "data_offset": 2048, 00:09:53.480 "data_size": 63488 00:09:53.480 }, 00:09:53.480 { 00:09:53.480 "name": "BaseBdev3", 00:09:53.480 "uuid": "9de0a78b-d84f-5612-bdf1-9be7487e0e2b", 00:09:53.480 "is_configured": true, 00:09:53.480 "data_offset": 2048, 00:09:53.480 "data_size": 63488 00:09:53.480 } 00:09:53.480 ] 00:09:53.480 }' 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:53.480 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.737 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:53.737 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.737 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.737 [2024-11-20 15:57:51.851426] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:53.737 [2024-11-20 15:57:51.851454] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:53.737 [2024-11-20 15:57:51.854531] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:53.737 [2024-11-20 15:57:51.854579] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:53.737 [2024-11-20 15:57:51.854614] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:53.737 [2024-11-20 15:57:51.854623] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:09:53.737 { 00:09:53.737 "results": [ 00:09:53.737 { 00:09:53.737 "job": "raid_bdev1", 00:09:53.737 "core_mask": "0x1", 00:09:53.737 "workload": "randrw", 00:09:53.737 "percentage": 50, 00:09:53.737 "status": "finished", 00:09:53.737 "queue_depth": 1, 00:09:53.737 "io_size": 131072, 00:09:53.737 "runtime": 1.24812, 00:09:53.737 "iops": 14875.973464089991, 00:09:53.737 "mibps": 1859.496683011249, 00:09:53.737 "io_failed": 1, 00:09:53.737 "io_timeout": 0, 00:09:53.737 "avg_latency_us": 91.50619693103104, 00:09:53.737 "min_latency_us": 33.673846153846156, 00:09:53.737 "max_latency_us": 1688.8123076923077 00:09:53.737 } 00:09:53.737 ], 00:09:53.737 "core_count": 1 00:09:53.737 } 00:09:53.737 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.737 15:57:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 63751 00:09:53.737 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 63751 ']' 00:09:53.737 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 63751 00:09:53.737 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:09:53.737 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:53.737 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 63751 00:09:53.737 killing process with pid 63751 00:09:53.737 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:53.737 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:53.737 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 63751' 00:09:53.737 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 63751 00:09:53.737 [2024-11-20 15:57:51.881082] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:53.737 15:57:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 63751 00:09:53.995 [2024-11-20 15:57:52.022804] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:54.561 15:57:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.QaujsBFrft 00:09:54.561 15:57:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:54.561 15:57:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:54.561 15:57:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.80 00:09:54.561 15:57:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:09:54.561 15:57:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:54.561 15:57:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:54.561 15:57:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.80 != \0\.\0\0 ]] 00:09:54.561 00:09:54.561 real 0m3.646s 00:09:54.561 user 0m4.303s 00:09:54.561 sys 0m0.407s 00:09:54.561 ************************************ 00:09:54.561 END TEST raid_read_error_test 00:09:54.561 ************************************ 00:09:54.561 15:57:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:54.561 15:57:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.819 15:57:52 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 3 write 00:09:54.819 15:57:52 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:54.819 15:57:52 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:54.819 15:57:52 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:54.819 ************************************ 00:09:54.819 START TEST raid_write_error_test 00:09:54.819 ************************************ 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 3 write 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.cAtxq1uWpR 00:09:54.819 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=63885 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 63885 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 63885 ']' 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.819 15:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:54.819 [2024-11-20 15:57:52.900003] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:09:54.819 [2024-11-20 15:57:52.900260] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid63885 ] 00:09:54.819 [2024-11-20 15:57:53.061605] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:55.077 [2024-11-20 15:57:53.161136] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:55.077 [2024-11-20 15:57:53.296439] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:55.077 [2024-11-20 15:57:53.296472] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:55.642 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:55.642 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:55.642 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:55.642 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:55.642 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.642 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.642 BaseBdev1_malloc 00:09:55.642 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.643 true 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.643 [2024-11-20 15:57:53.781913] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:55.643 [2024-11-20 15:57:53.781966] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:55.643 [2024-11-20 15:57:53.781985] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:55.643 [2024-11-20 15:57:53.781996] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:55.643 [2024-11-20 15:57:53.784104] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:55.643 [2024-11-20 15:57:53.784142] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:55.643 BaseBdev1 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.643 BaseBdev2_malloc 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.643 true 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.643 [2024-11-20 15:57:53.825446] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:55.643 [2024-11-20 15:57:53.825495] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:55.643 [2024-11-20 15:57:53.825512] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:55.643 [2024-11-20 15:57:53.825523] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:55.643 [2024-11-20 15:57:53.827617] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:55.643 [2024-11-20 15:57:53.827777] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:55.643 BaseBdev2 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.643 BaseBdev3_malloc 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.643 true 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.643 [2024-11-20 15:57:53.880476] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:55.643 [2024-11-20 15:57:53.880640] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:55.643 [2024-11-20 15:57:53.880663] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:55.643 [2024-11-20 15:57:53.880692] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:55.643 [2024-11-20 15:57:53.882800] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:55.643 [2024-11-20 15:57:53.882834] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:55.643 BaseBdev3 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.643 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.643 [2024-11-20 15:57:53.888552] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:55.643 [2024-11-20 15:57:53.890360] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:55.643 [2024-11-20 15:57:53.890434] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:55.643 [2024-11-20 15:57:53.890618] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:09:55.643 [2024-11-20 15:57:53.890631] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:55.643 [2024-11-20 15:57:53.890900] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:09:55.643 [2024-11-20 15:57:53.891040] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:09:55.900 [2024-11-20 15:57:53.891126] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:09:55.900 [2024-11-20 15:57:53.891264] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:55.900 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.900 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:09:55.900 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:55.900 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:55.900 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:55.900 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:55.900 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:55.900 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:55.900 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:55.900 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:55.900 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:55.900 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.901 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:55.901 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.901 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.901 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.901 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:55.901 "name": "raid_bdev1", 00:09:55.901 "uuid": "d0d5f055-02cc-4930-8105-ddacd43b1f42", 00:09:55.901 "strip_size_kb": 64, 00:09:55.901 "state": "online", 00:09:55.901 "raid_level": "raid0", 00:09:55.901 "superblock": true, 00:09:55.901 "num_base_bdevs": 3, 00:09:55.901 "num_base_bdevs_discovered": 3, 00:09:55.901 "num_base_bdevs_operational": 3, 00:09:55.901 "base_bdevs_list": [ 00:09:55.901 { 00:09:55.901 "name": "BaseBdev1", 00:09:55.901 "uuid": "df73e5c2-5a1f-553f-8eb0-875408dc26a1", 00:09:55.901 "is_configured": true, 00:09:55.901 "data_offset": 2048, 00:09:55.901 "data_size": 63488 00:09:55.901 }, 00:09:55.901 { 00:09:55.901 "name": "BaseBdev2", 00:09:55.901 "uuid": "a5d43177-973a-5a22-9f20-0a502f37c5d3", 00:09:55.901 "is_configured": true, 00:09:55.901 "data_offset": 2048, 00:09:55.901 "data_size": 63488 00:09:55.901 }, 00:09:55.901 { 00:09:55.901 "name": "BaseBdev3", 00:09:55.901 "uuid": "185b5109-0547-5607-b273-4f816d87a60e", 00:09:55.901 "is_configured": true, 00:09:55.901 "data_offset": 2048, 00:09:55.901 "data_size": 63488 00:09:55.901 } 00:09:55.901 ] 00:09:55.901 }' 00:09:55.901 15:57:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:55.901 15:57:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.157 15:57:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:56.157 15:57:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:56.157 [2024-11-20 15:57:54.365567] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:57.088 "name": "raid_bdev1", 00:09:57.088 "uuid": "d0d5f055-02cc-4930-8105-ddacd43b1f42", 00:09:57.088 "strip_size_kb": 64, 00:09:57.088 "state": "online", 00:09:57.088 "raid_level": "raid0", 00:09:57.088 "superblock": true, 00:09:57.088 "num_base_bdevs": 3, 00:09:57.088 "num_base_bdevs_discovered": 3, 00:09:57.088 "num_base_bdevs_operational": 3, 00:09:57.088 "base_bdevs_list": [ 00:09:57.088 { 00:09:57.088 "name": "BaseBdev1", 00:09:57.088 "uuid": "df73e5c2-5a1f-553f-8eb0-875408dc26a1", 00:09:57.088 "is_configured": true, 00:09:57.088 "data_offset": 2048, 00:09:57.088 "data_size": 63488 00:09:57.088 }, 00:09:57.088 { 00:09:57.088 "name": "BaseBdev2", 00:09:57.088 "uuid": "a5d43177-973a-5a22-9f20-0a502f37c5d3", 00:09:57.088 "is_configured": true, 00:09:57.088 "data_offset": 2048, 00:09:57.088 "data_size": 63488 00:09:57.088 }, 00:09:57.088 { 00:09:57.088 "name": "BaseBdev3", 00:09:57.088 "uuid": "185b5109-0547-5607-b273-4f816d87a60e", 00:09:57.088 "is_configured": true, 00:09:57.088 "data_offset": 2048, 00:09:57.088 "data_size": 63488 00:09:57.088 } 00:09:57.088 ] 00:09:57.088 }' 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:57.088 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.351 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:57.351 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.351 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.351 [2024-11-20 15:57:55.574691] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:57.351 [2024-11-20 15:57:55.574716] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:57.351 [2024-11-20 15:57:55.577194] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:57.351 [2024-11-20 15:57:55.577339] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:57.351 [2024-11-20 15:57:55.577378] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:57.351 [2024-11-20 15:57:55.577386] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:09:57.351 { 00:09:57.351 "results": [ 00:09:57.351 { 00:09:57.351 "job": "raid_bdev1", 00:09:57.351 "core_mask": "0x1", 00:09:57.351 "workload": "randrw", 00:09:57.351 "percentage": 50, 00:09:57.351 "status": "finished", 00:09:57.351 "queue_depth": 1, 00:09:57.351 "io_size": 131072, 00:09:57.351 "runtime": 1.207281, 00:09:57.351 "iops": 15889.424251686227, 00:09:57.351 "mibps": 1986.1780314607784, 00:09:57.351 "io_failed": 1, 00:09:57.351 "io_timeout": 0, 00:09:57.351 "avg_latency_us": 85.76828126002438, 00:09:57.351 "min_latency_us": 15.95076923076923, 00:09:57.351 "max_latency_us": 1310.72 00:09:57.351 } 00:09:57.351 ], 00:09:57.351 "core_count": 1 00:09:57.351 } 00:09:57.351 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.351 15:57:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 63885 00:09:57.351 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 63885 ']' 00:09:57.351 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 63885 00:09:57.351 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:09:57.351 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:57.351 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 63885 00:09:57.616 killing process with pid 63885 00:09:57.616 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:57.616 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:57.616 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 63885' 00:09:57.616 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 63885 00:09:57.616 [2024-11-20 15:57:55.606386] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:57.616 15:57:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 63885 00:09:57.616 [2024-11-20 15:57:55.720379] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:58.181 15:57:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:58.181 15:57:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.cAtxq1uWpR 00:09:58.181 15:57:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:58.181 15:57:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.83 00:09:58.181 15:57:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:09:58.181 15:57:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:58.181 15:57:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:58.181 15:57:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.83 != \0\.\0\0 ]] 00:09:58.181 00:09:58.181 real 0m3.495s 00:09:58.181 user 0m4.285s 00:09:58.181 sys 0m0.358s 00:09:58.181 15:57:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:58.181 15:57:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.181 ************************************ 00:09:58.181 END TEST raid_write_error_test 00:09:58.181 ************************************ 00:09:58.181 15:57:56 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:58.181 15:57:56 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 3 false 00:09:58.181 15:57:56 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:58.181 15:57:56 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:58.181 15:57:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:58.181 ************************************ 00:09:58.181 START TEST raid_state_function_test 00:09:58.181 ************************************ 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 3 false 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:58.181 Process raid pid: 64018 00:09:58.181 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=64018 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 64018' 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 64018 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 64018 ']' 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:58.181 15:57:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.439 [2024-11-20 15:57:56.443449] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:09:58.439 [2024-11-20 15:57:56.443805] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:58.439 [2024-11-20 15:57:56.604964] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:58.696 [2024-11-20 15:57:56.703541] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:58.696 [2024-11-20 15:57:56.838834] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:58.696 [2024-11-20 15:57:56.838858] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.260 [2024-11-20 15:57:57.284129] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:59.260 [2024-11-20 15:57:57.284178] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:59.260 [2024-11-20 15:57:57.284188] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:59.260 [2024-11-20 15:57:57.284198] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:59.260 [2024-11-20 15:57:57.284205] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:59.260 [2024-11-20 15:57:57.284213] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:59.260 "name": "Existed_Raid", 00:09:59.260 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.260 "strip_size_kb": 64, 00:09:59.260 "state": "configuring", 00:09:59.260 "raid_level": "concat", 00:09:59.260 "superblock": false, 00:09:59.260 "num_base_bdevs": 3, 00:09:59.260 "num_base_bdevs_discovered": 0, 00:09:59.260 "num_base_bdevs_operational": 3, 00:09:59.260 "base_bdevs_list": [ 00:09:59.260 { 00:09:59.260 "name": "BaseBdev1", 00:09:59.260 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.260 "is_configured": false, 00:09:59.260 "data_offset": 0, 00:09:59.260 "data_size": 0 00:09:59.260 }, 00:09:59.260 { 00:09:59.260 "name": "BaseBdev2", 00:09:59.260 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.260 "is_configured": false, 00:09:59.260 "data_offset": 0, 00:09:59.260 "data_size": 0 00:09:59.260 }, 00:09:59.260 { 00:09:59.260 "name": "BaseBdev3", 00:09:59.260 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.260 "is_configured": false, 00:09:59.260 "data_offset": 0, 00:09:59.260 "data_size": 0 00:09:59.260 } 00:09:59.260 ] 00:09:59.260 }' 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:59.260 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.518 [2024-11-20 15:57:57.616130] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:59.518 [2024-11-20 15:57:57.616159] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.518 [2024-11-20 15:57:57.624143] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:59.518 [2024-11-20 15:57:57.624179] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:59.518 [2024-11-20 15:57:57.624187] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:59.518 [2024-11-20 15:57:57.624196] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:59.518 [2024-11-20 15:57:57.624202] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:59.518 [2024-11-20 15:57:57.624211] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.518 [2024-11-20 15:57:57.656181] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:59.518 BaseBdev1 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.518 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.518 [ 00:09:59.518 { 00:09:59.518 "name": "BaseBdev1", 00:09:59.518 "aliases": [ 00:09:59.518 "ec508913-3f84-4fce-91f4-4b8ac96cab29" 00:09:59.518 ], 00:09:59.518 "product_name": "Malloc disk", 00:09:59.518 "block_size": 512, 00:09:59.518 "num_blocks": 65536, 00:09:59.518 "uuid": "ec508913-3f84-4fce-91f4-4b8ac96cab29", 00:09:59.518 "assigned_rate_limits": { 00:09:59.518 "rw_ios_per_sec": 0, 00:09:59.518 "rw_mbytes_per_sec": 0, 00:09:59.518 "r_mbytes_per_sec": 0, 00:09:59.518 "w_mbytes_per_sec": 0 00:09:59.518 }, 00:09:59.518 "claimed": true, 00:09:59.518 "claim_type": "exclusive_write", 00:09:59.518 "zoned": false, 00:09:59.518 "supported_io_types": { 00:09:59.518 "read": true, 00:09:59.518 "write": true, 00:09:59.518 "unmap": true, 00:09:59.518 "flush": true, 00:09:59.518 "reset": true, 00:09:59.518 "nvme_admin": false, 00:09:59.518 "nvme_io": false, 00:09:59.518 "nvme_io_md": false, 00:09:59.518 "write_zeroes": true, 00:09:59.518 "zcopy": true, 00:09:59.518 "get_zone_info": false, 00:09:59.518 "zone_management": false, 00:09:59.518 "zone_append": false, 00:09:59.518 "compare": false, 00:09:59.518 "compare_and_write": false, 00:09:59.518 "abort": true, 00:09:59.518 "seek_hole": false, 00:09:59.518 "seek_data": false, 00:09:59.518 "copy": true, 00:09:59.518 "nvme_iov_md": false 00:09:59.518 }, 00:09:59.518 "memory_domains": [ 00:09:59.518 { 00:09:59.519 "dma_device_id": "system", 00:09:59.519 "dma_device_type": 1 00:09:59.519 }, 00:09:59.519 { 00:09:59.519 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:59.519 "dma_device_type": 2 00:09:59.519 } 00:09:59.519 ], 00:09:59.519 "driver_specific": {} 00:09:59.519 } 00:09:59.519 ] 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:59.519 "name": "Existed_Raid", 00:09:59.519 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.519 "strip_size_kb": 64, 00:09:59.519 "state": "configuring", 00:09:59.519 "raid_level": "concat", 00:09:59.519 "superblock": false, 00:09:59.519 "num_base_bdevs": 3, 00:09:59.519 "num_base_bdevs_discovered": 1, 00:09:59.519 "num_base_bdevs_operational": 3, 00:09:59.519 "base_bdevs_list": [ 00:09:59.519 { 00:09:59.519 "name": "BaseBdev1", 00:09:59.519 "uuid": "ec508913-3f84-4fce-91f4-4b8ac96cab29", 00:09:59.519 "is_configured": true, 00:09:59.519 "data_offset": 0, 00:09:59.519 "data_size": 65536 00:09:59.519 }, 00:09:59.519 { 00:09:59.519 "name": "BaseBdev2", 00:09:59.519 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.519 "is_configured": false, 00:09:59.519 "data_offset": 0, 00:09:59.519 "data_size": 0 00:09:59.519 }, 00:09:59.519 { 00:09:59.519 "name": "BaseBdev3", 00:09:59.519 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.519 "is_configured": false, 00:09:59.519 "data_offset": 0, 00:09:59.519 "data_size": 0 00:09:59.519 } 00:09:59.519 ] 00:09:59.519 }' 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:59.519 15:57:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.776 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:59.776 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.776 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.034 [2024-11-20 15:57:58.028296] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:00.034 [2024-11-20 15:57:58.028336] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.034 [2024-11-20 15:57:58.036348] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:00.034 [2024-11-20 15:57:58.038210] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:00.034 [2024-11-20 15:57:58.038249] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:00.034 [2024-11-20 15:57:58.038258] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:00.034 [2024-11-20 15:57:58.038267] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:00.034 "name": "Existed_Raid", 00:10:00.034 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:00.034 "strip_size_kb": 64, 00:10:00.034 "state": "configuring", 00:10:00.034 "raid_level": "concat", 00:10:00.034 "superblock": false, 00:10:00.034 "num_base_bdevs": 3, 00:10:00.034 "num_base_bdevs_discovered": 1, 00:10:00.034 "num_base_bdevs_operational": 3, 00:10:00.034 "base_bdevs_list": [ 00:10:00.034 { 00:10:00.034 "name": "BaseBdev1", 00:10:00.034 "uuid": "ec508913-3f84-4fce-91f4-4b8ac96cab29", 00:10:00.034 "is_configured": true, 00:10:00.034 "data_offset": 0, 00:10:00.034 "data_size": 65536 00:10:00.034 }, 00:10:00.034 { 00:10:00.034 "name": "BaseBdev2", 00:10:00.034 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:00.034 "is_configured": false, 00:10:00.034 "data_offset": 0, 00:10:00.034 "data_size": 0 00:10:00.034 }, 00:10:00.034 { 00:10:00.034 "name": "BaseBdev3", 00:10:00.034 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:00.034 "is_configured": false, 00:10:00.034 "data_offset": 0, 00:10:00.034 "data_size": 0 00:10:00.034 } 00:10:00.034 ] 00:10:00.034 }' 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:00.034 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.292 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:00.292 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.292 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.292 [2024-11-20 15:57:58.370543] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:00.292 BaseBdev2 00:10:00.292 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.292 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:00.292 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:00.292 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:00.292 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:00.292 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:00.292 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:00.292 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:00.292 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.292 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.292 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.292 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:00.292 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.292 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.292 [ 00:10:00.292 { 00:10:00.292 "name": "BaseBdev2", 00:10:00.292 "aliases": [ 00:10:00.292 "ab082fb6-3707-4e47-901b-c18a959c4957" 00:10:00.292 ], 00:10:00.292 "product_name": "Malloc disk", 00:10:00.292 "block_size": 512, 00:10:00.292 "num_blocks": 65536, 00:10:00.292 "uuid": "ab082fb6-3707-4e47-901b-c18a959c4957", 00:10:00.292 "assigned_rate_limits": { 00:10:00.292 "rw_ios_per_sec": 0, 00:10:00.292 "rw_mbytes_per_sec": 0, 00:10:00.292 "r_mbytes_per_sec": 0, 00:10:00.292 "w_mbytes_per_sec": 0 00:10:00.292 }, 00:10:00.292 "claimed": true, 00:10:00.292 "claim_type": "exclusive_write", 00:10:00.292 "zoned": false, 00:10:00.292 "supported_io_types": { 00:10:00.292 "read": true, 00:10:00.292 "write": true, 00:10:00.292 "unmap": true, 00:10:00.292 "flush": true, 00:10:00.292 "reset": true, 00:10:00.292 "nvme_admin": false, 00:10:00.292 "nvme_io": false, 00:10:00.292 "nvme_io_md": false, 00:10:00.292 "write_zeroes": true, 00:10:00.292 "zcopy": true, 00:10:00.292 "get_zone_info": false, 00:10:00.292 "zone_management": false, 00:10:00.292 "zone_append": false, 00:10:00.292 "compare": false, 00:10:00.292 "compare_and_write": false, 00:10:00.293 "abort": true, 00:10:00.293 "seek_hole": false, 00:10:00.293 "seek_data": false, 00:10:00.293 "copy": true, 00:10:00.293 "nvme_iov_md": false 00:10:00.293 }, 00:10:00.293 "memory_domains": [ 00:10:00.293 { 00:10:00.293 "dma_device_id": "system", 00:10:00.293 "dma_device_type": 1 00:10:00.293 }, 00:10:00.293 { 00:10:00.293 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:00.293 "dma_device_type": 2 00:10:00.293 } 00:10:00.293 ], 00:10:00.293 "driver_specific": {} 00:10:00.293 } 00:10:00.293 ] 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:00.293 "name": "Existed_Raid", 00:10:00.293 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:00.293 "strip_size_kb": 64, 00:10:00.293 "state": "configuring", 00:10:00.293 "raid_level": "concat", 00:10:00.293 "superblock": false, 00:10:00.293 "num_base_bdevs": 3, 00:10:00.293 "num_base_bdevs_discovered": 2, 00:10:00.293 "num_base_bdevs_operational": 3, 00:10:00.293 "base_bdevs_list": [ 00:10:00.293 { 00:10:00.293 "name": "BaseBdev1", 00:10:00.293 "uuid": "ec508913-3f84-4fce-91f4-4b8ac96cab29", 00:10:00.293 "is_configured": true, 00:10:00.293 "data_offset": 0, 00:10:00.293 "data_size": 65536 00:10:00.293 }, 00:10:00.293 { 00:10:00.293 "name": "BaseBdev2", 00:10:00.293 "uuid": "ab082fb6-3707-4e47-901b-c18a959c4957", 00:10:00.293 "is_configured": true, 00:10:00.293 "data_offset": 0, 00:10:00.293 "data_size": 65536 00:10:00.293 }, 00:10:00.293 { 00:10:00.293 "name": "BaseBdev3", 00:10:00.293 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:00.293 "is_configured": false, 00:10:00.293 "data_offset": 0, 00:10:00.293 "data_size": 0 00:10:00.293 } 00:10:00.293 ] 00:10:00.293 }' 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:00.293 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.550 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.551 [2024-11-20 15:57:58.750745] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:00.551 [2024-11-20 15:57:58.750907] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:10:00.551 [2024-11-20 15:57:58.750945] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:10:00.551 [2024-11-20 15:57:58.751343] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:10:00.551 [2024-11-20 15:57:58.751580] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:10:00.551 [2024-11-20 15:57:58.751611] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:10:00.551 [2024-11-20 15:57:58.751912] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:00.551 BaseBdev3 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.551 [ 00:10:00.551 { 00:10:00.551 "name": "BaseBdev3", 00:10:00.551 "aliases": [ 00:10:00.551 "19f722db-64ed-4a4e-9d2b-0fb75704e743" 00:10:00.551 ], 00:10:00.551 "product_name": "Malloc disk", 00:10:00.551 "block_size": 512, 00:10:00.551 "num_blocks": 65536, 00:10:00.551 "uuid": "19f722db-64ed-4a4e-9d2b-0fb75704e743", 00:10:00.551 "assigned_rate_limits": { 00:10:00.551 "rw_ios_per_sec": 0, 00:10:00.551 "rw_mbytes_per_sec": 0, 00:10:00.551 "r_mbytes_per_sec": 0, 00:10:00.551 "w_mbytes_per_sec": 0 00:10:00.551 }, 00:10:00.551 "claimed": true, 00:10:00.551 "claim_type": "exclusive_write", 00:10:00.551 "zoned": false, 00:10:00.551 "supported_io_types": { 00:10:00.551 "read": true, 00:10:00.551 "write": true, 00:10:00.551 "unmap": true, 00:10:00.551 "flush": true, 00:10:00.551 "reset": true, 00:10:00.551 "nvme_admin": false, 00:10:00.551 "nvme_io": false, 00:10:00.551 "nvme_io_md": false, 00:10:00.551 "write_zeroes": true, 00:10:00.551 "zcopy": true, 00:10:00.551 "get_zone_info": false, 00:10:00.551 "zone_management": false, 00:10:00.551 "zone_append": false, 00:10:00.551 "compare": false, 00:10:00.551 "compare_and_write": false, 00:10:00.551 "abort": true, 00:10:00.551 "seek_hole": false, 00:10:00.551 "seek_data": false, 00:10:00.551 "copy": true, 00:10:00.551 "nvme_iov_md": false 00:10:00.551 }, 00:10:00.551 "memory_domains": [ 00:10:00.551 { 00:10:00.551 "dma_device_id": "system", 00:10:00.551 "dma_device_type": 1 00:10:00.551 }, 00:10:00.551 { 00:10:00.551 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:00.551 "dma_device_type": 2 00:10:00.551 } 00:10:00.551 ], 00:10:00.551 "driver_specific": {} 00:10:00.551 } 00:10:00.551 ] 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:00.551 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.808 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:00.808 "name": "Existed_Raid", 00:10:00.808 "uuid": "2783297e-745c-41ff-a714-b91fbcd302bb", 00:10:00.808 "strip_size_kb": 64, 00:10:00.808 "state": "online", 00:10:00.808 "raid_level": "concat", 00:10:00.808 "superblock": false, 00:10:00.808 "num_base_bdevs": 3, 00:10:00.808 "num_base_bdevs_discovered": 3, 00:10:00.808 "num_base_bdevs_operational": 3, 00:10:00.808 "base_bdevs_list": [ 00:10:00.808 { 00:10:00.808 "name": "BaseBdev1", 00:10:00.808 "uuid": "ec508913-3f84-4fce-91f4-4b8ac96cab29", 00:10:00.808 "is_configured": true, 00:10:00.808 "data_offset": 0, 00:10:00.808 "data_size": 65536 00:10:00.808 }, 00:10:00.808 { 00:10:00.808 "name": "BaseBdev2", 00:10:00.808 "uuid": "ab082fb6-3707-4e47-901b-c18a959c4957", 00:10:00.808 "is_configured": true, 00:10:00.808 "data_offset": 0, 00:10:00.808 "data_size": 65536 00:10:00.808 }, 00:10:00.808 { 00:10:00.808 "name": "BaseBdev3", 00:10:00.808 "uuid": "19f722db-64ed-4a4e-9d2b-0fb75704e743", 00:10:00.808 "is_configured": true, 00:10:00.808 "data_offset": 0, 00:10:00.808 "data_size": 65536 00:10:00.808 } 00:10:00.808 ] 00:10:00.808 }' 00:10:00.808 15:57:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:00.808 15:57:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.066 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:01.066 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:01.066 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:01.066 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:01.066 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:01.066 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:01.066 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:01.066 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:01.066 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.066 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.066 [2024-11-20 15:57:59.071170] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:01.066 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.066 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:01.066 "name": "Existed_Raid", 00:10:01.066 "aliases": [ 00:10:01.066 "2783297e-745c-41ff-a714-b91fbcd302bb" 00:10:01.066 ], 00:10:01.066 "product_name": "Raid Volume", 00:10:01.066 "block_size": 512, 00:10:01.066 "num_blocks": 196608, 00:10:01.066 "uuid": "2783297e-745c-41ff-a714-b91fbcd302bb", 00:10:01.066 "assigned_rate_limits": { 00:10:01.066 "rw_ios_per_sec": 0, 00:10:01.066 "rw_mbytes_per_sec": 0, 00:10:01.066 "r_mbytes_per_sec": 0, 00:10:01.066 "w_mbytes_per_sec": 0 00:10:01.066 }, 00:10:01.066 "claimed": false, 00:10:01.066 "zoned": false, 00:10:01.066 "supported_io_types": { 00:10:01.066 "read": true, 00:10:01.066 "write": true, 00:10:01.066 "unmap": true, 00:10:01.067 "flush": true, 00:10:01.067 "reset": true, 00:10:01.067 "nvme_admin": false, 00:10:01.067 "nvme_io": false, 00:10:01.067 "nvme_io_md": false, 00:10:01.067 "write_zeroes": true, 00:10:01.067 "zcopy": false, 00:10:01.067 "get_zone_info": false, 00:10:01.067 "zone_management": false, 00:10:01.067 "zone_append": false, 00:10:01.067 "compare": false, 00:10:01.067 "compare_and_write": false, 00:10:01.067 "abort": false, 00:10:01.067 "seek_hole": false, 00:10:01.067 "seek_data": false, 00:10:01.067 "copy": false, 00:10:01.067 "nvme_iov_md": false 00:10:01.067 }, 00:10:01.067 "memory_domains": [ 00:10:01.067 { 00:10:01.067 "dma_device_id": "system", 00:10:01.067 "dma_device_type": 1 00:10:01.067 }, 00:10:01.067 { 00:10:01.067 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.067 "dma_device_type": 2 00:10:01.067 }, 00:10:01.067 { 00:10:01.067 "dma_device_id": "system", 00:10:01.067 "dma_device_type": 1 00:10:01.067 }, 00:10:01.067 { 00:10:01.067 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.067 "dma_device_type": 2 00:10:01.067 }, 00:10:01.067 { 00:10:01.067 "dma_device_id": "system", 00:10:01.067 "dma_device_type": 1 00:10:01.067 }, 00:10:01.067 { 00:10:01.067 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.067 "dma_device_type": 2 00:10:01.067 } 00:10:01.067 ], 00:10:01.067 "driver_specific": { 00:10:01.067 "raid": { 00:10:01.067 "uuid": "2783297e-745c-41ff-a714-b91fbcd302bb", 00:10:01.067 "strip_size_kb": 64, 00:10:01.067 "state": "online", 00:10:01.067 "raid_level": "concat", 00:10:01.067 "superblock": false, 00:10:01.067 "num_base_bdevs": 3, 00:10:01.067 "num_base_bdevs_discovered": 3, 00:10:01.067 "num_base_bdevs_operational": 3, 00:10:01.067 "base_bdevs_list": [ 00:10:01.067 { 00:10:01.067 "name": "BaseBdev1", 00:10:01.067 "uuid": "ec508913-3f84-4fce-91f4-4b8ac96cab29", 00:10:01.067 "is_configured": true, 00:10:01.067 "data_offset": 0, 00:10:01.067 "data_size": 65536 00:10:01.067 }, 00:10:01.067 { 00:10:01.067 "name": "BaseBdev2", 00:10:01.067 "uuid": "ab082fb6-3707-4e47-901b-c18a959c4957", 00:10:01.067 "is_configured": true, 00:10:01.067 "data_offset": 0, 00:10:01.067 "data_size": 65536 00:10:01.067 }, 00:10:01.067 { 00:10:01.067 "name": "BaseBdev3", 00:10:01.067 "uuid": "19f722db-64ed-4a4e-9d2b-0fb75704e743", 00:10:01.067 "is_configured": true, 00:10:01.067 "data_offset": 0, 00:10:01.067 "data_size": 65536 00:10:01.067 } 00:10:01.067 ] 00:10:01.067 } 00:10:01.067 } 00:10:01.067 }' 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:01.067 BaseBdev2 00:10:01.067 BaseBdev3' 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.067 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.067 [2024-11-20 15:57:59.262929] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:01.067 [2024-11-20 15:57:59.262952] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:01.067 [2024-11-20 15:57:59.262999] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:01.325 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.325 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:01.325 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:10:01.325 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:01.326 "name": "Existed_Raid", 00:10:01.326 "uuid": "2783297e-745c-41ff-a714-b91fbcd302bb", 00:10:01.326 "strip_size_kb": 64, 00:10:01.326 "state": "offline", 00:10:01.326 "raid_level": "concat", 00:10:01.326 "superblock": false, 00:10:01.326 "num_base_bdevs": 3, 00:10:01.326 "num_base_bdevs_discovered": 2, 00:10:01.326 "num_base_bdevs_operational": 2, 00:10:01.326 "base_bdevs_list": [ 00:10:01.326 { 00:10:01.326 "name": null, 00:10:01.326 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:01.326 "is_configured": false, 00:10:01.326 "data_offset": 0, 00:10:01.326 "data_size": 65536 00:10:01.326 }, 00:10:01.326 { 00:10:01.326 "name": "BaseBdev2", 00:10:01.326 "uuid": "ab082fb6-3707-4e47-901b-c18a959c4957", 00:10:01.326 "is_configured": true, 00:10:01.326 "data_offset": 0, 00:10:01.326 "data_size": 65536 00:10:01.326 }, 00:10:01.326 { 00:10:01.326 "name": "BaseBdev3", 00:10:01.326 "uuid": "19f722db-64ed-4a4e-9d2b-0fb75704e743", 00:10:01.326 "is_configured": true, 00:10:01.326 "data_offset": 0, 00:10:01.326 "data_size": 65536 00:10:01.326 } 00:10:01.326 ] 00:10:01.326 }' 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:01.326 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.584 [2024-11-20 15:57:59.665759] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.584 [2024-11-20 15:57:59.751664] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:01.584 [2024-11-20 15:57:59.751709] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.584 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.843 BaseBdev2 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.843 [ 00:10:01.843 { 00:10:01.843 "name": "BaseBdev2", 00:10:01.843 "aliases": [ 00:10:01.843 "e7de92a6-90de-4e89-b52e-b3c949f8f0dd" 00:10:01.843 ], 00:10:01.843 "product_name": "Malloc disk", 00:10:01.843 "block_size": 512, 00:10:01.843 "num_blocks": 65536, 00:10:01.843 "uuid": "e7de92a6-90de-4e89-b52e-b3c949f8f0dd", 00:10:01.843 "assigned_rate_limits": { 00:10:01.843 "rw_ios_per_sec": 0, 00:10:01.843 "rw_mbytes_per_sec": 0, 00:10:01.843 "r_mbytes_per_sec": 0, 00:10:01.843 "w_mbytes_per_sec": 0 00:10:01.843 }, 00:10:01.843 "claimed": false, 00:10:01.843 "zoned": false, 00:10:01.843 "supported_io_types": { 00:10:01.843 "read": true, 00:10:01.843 "write": true, 00:10:01.843 "unmap": true, 00:10:01.843 "flush": true, 00:10:01.843 "reset": true, 00:10:01.843 "nvme_admin": false, 00:10:01.843 "nvme_io": false, 00:10:01.843 "nvme_io_md": false, 00:10:01.843 "write_zeroes": true, 00:10:01.843 "zcopy": true, 00:10:01.843 "get_zone_info": false, 00:10:01.843 "zone_management": false, 00:10:01.843 "zone_append": false, 00:10:01.843 "compare": false, 00:10:01.843 "compare_and_write": false, 00:10:01.843 "abort": true, 00:10:01.843 "seek_hole": false, 00:10:01.843 "seek_data": false, 00:10:01.843 "copy": true, 00:10:01.843 "nvme_iov_md": false 00:10:01.843 }, 00:10:01.843 "memory_domains": [ 00:10:01.843 { 00:10:01.843 "dma_device_id": "system", 00:10:01.843 "dma_device_type": 1 00:10:01.843 }, 00:10:01.843 { 00:10:01.843 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.843 "dma_device_type": 2 00:10:01.843 } 00:10:01.843 ], 00:10:01.843 "driver_specific": {} 00:10:01.843 } 00:10:01.843 ] 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.843 BaseBdev3 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:01.843 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.844 [ 00:10:01.844 { 00:10:01.844 "name": "BaseBdev3", 00:10:01.844 "aliases": [ 00:10:01.844 "73885ed7-0d71-44d2-a3f9-239d19e04710" 00:10:01.844 ], 00:10:01.844 "product_name": "Malloc disk", 00:10:01.844 "block_size": 512, 00:10:01.844 "num_blocks": 65536, 00:10:01.844 "uuid": "73885ed7-0d71-44d2-a3f9-239d19e04710", 00:10:01.844 "assigned_rate_limits": { 00:10:01.844 "rw_ios_per_sec": 0, 00:10:01.844 "rw_mbytes_per_sec": 0, 00:10:01.844 "r_mbytes_per_sec": 0, 00:10:01.844 "w_mbytes_per_sec": 0 00:10:01.844 }, 00:10:01.844 "claimed": false, 00:10:01.844 "zoned": false, 00:10:01.844 "supported_io_types": { 00:10:01.844 "read": true, 00:10:01.844 "write": true, 00:10:01.844 "unmap": true, 00:10:01.844 "flush": true, 00:10:01.844 "reset": true, 00:10:01.844 "nvme_admin": false, 00:10:01.844 "nvme_io": false, 00:10:01.844 "nvme_io_md": false, 00:10:01.844 "write_zeroes": true, 00:10:01.844 "zcopy": true, 00:10:01.844 "get_zone_info": false, 00:10:01.844 "zone_management": false, 00:10:01.844 "zone_append": false, 00:10:01.844 "compare": false, 00:10:01.844 "compare_and_write": false, 00:10:01.844 "abort": true, 00:10:01.844 "seek_hole": false, 00:10:01.844 "seek_data": false, 00:10:01.844 "copy": true, 00:10:01.844 "nvme_iov_md": false 00:10:01.844 }, 00:10:01.844 "memory_domains": [ 00:10:01.844 { 00:10:01.844 "dma_device_id": "system", 00:10:01.844 "dma_device_type": 1 00:10:01.844 }, 00:10:01.844 { 00:10:01.844 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.844 "dma_device_type": 2 00:10:01.844 } 00:10:01.844 ], 00:10:01.844 "driver_specific": {} 00:10:01.844 } 00:10:01.844 ] 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.844 [2024-11-20 15:57:59.935419] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:01.844 [2024-11-20 15:57:59.935455] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:01.844 [2024-11-20 15:57:59.935472] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:01.844 [2024-11-20 15:57:59.936947] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:01.844 "name": "Existed_Raid", 00:10:01.844 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:01.844 "strip_size_kb": 64, 00:10:01.844 "state": "configuring", 00:10:01.844 "raid_level": "concat", 00:10:01.844 "superblock": false, 00:10:01.844 "num_base_bdevs": 3, 00:10:01.844 "num_base_bdevs_discovered": 2, 00:10:01.844 "num_base_bdevs_operational": 3, 00:10:01.844 "base_bdevs_list": [ 00:10:01.844 { 00:10:01.844 "name": "BaseBdev1", 00:10:01.844 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:01.844 "is_configured": false, 00:10:01.844 "data_offset": 0, 00:10:01.844 "data_size": 0 00:10:01.844 }, 00:10:01.844 { 00:10:01.844 "name": "BaseBdev2", 00:10:01.844 "uuid": "e7de92a6-90de-4e89-b52e-b3c949f8f0dd", 00:10:01.844 "is_configured": true, 00:10:01.844 "data_offset": 0, 00:10:01.844 "data_size": 65536 00:10:01.844 }, 00:10:01.844 { 00:10:01.844 "name": "BaseBdev3", 00:10:01.844 "uuid": "73885ed7-0d71-44d2-a3f9-239d19e04710", 00:10:01.844 "is_configured": true, 00:10:01.844 "data_offset": 0, 00:10:01.844 "data_size": 65536 00:10:01.844 } 00:10:01.844 ] 00:10:01.844 }' 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:01.844 15:57:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.102 [2024-11-20 15:58:00.247498] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:02.102 "name": "Existed_Raid", 00:10:02.102 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:02.102 "strip_size_kb": 64, 00:10:02.102 "state": "configuring", 00:10:02.102 "raid_level": "concat", 00:10:02.102 "superblock": false, 00:10:02.102 "num_base_bdevs": 3, 00:10:02.102 "num_base_bdevs_discovered": 1, 00:10:02.102 "num_base_bdevs_operational": 3, 00:10:02.102 "base_bdevs_list": [ 00:10:02.102 { 00:10:02.102 "name": "BaseBdev1", 00:10:02.102 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:02.102 "is_configured": false, 00:10:02.102 "data_offset": 0, 00:10:02.102 "data_size": 0 00:10:02.102 }, 00:10:02.102 { 00:10:02.102 "name": null, 00:10:02.102 "uuid": "e7de92a6-90de-4e89-b52e-b3c949f8f0dd", 00:10:02.102 "is_configured": false, 00:10:02.102 "data_offset": 0, 00:10:02.102 "data_size": 65536 00:10:02.102 }, 00:10:02.102 { 00:10:02.102 "name": "BaseBdev3", 00:10:02.102 "uuid": "73885ed7-0d71-44d2-a3f9-239d19e04710", 00:10:02.102 "is_configured": true, 00:10:02.102 "data_offset": 0, 00:10:02.102 "data_size": 65536 00:10:02.102 } 00:10:02.102 ] 00:10:02.102 }' 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:02.102 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.359 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.359 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:02.359 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.359 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.359 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.359 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:02.359 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:02.359 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.359 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.618 [2024-11-20 15:58:00.609520] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:02.618 BaseBdev1 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.618 [ 00:10:02.618 { 00:10:02.618 "name": "BaseBdev1", 00:10:02.618 "aliases": [ 00:10:02.618 "c7383938-46cd-4a21-a97d-80b55d8b7888" 00:10:02.618 ], 00:10:02.618 "product_name": "Malloc disk", 00:10:02.618 "block_size": 512, 00:10:02.618 "num_blocks": 65536, 00:10:02.618 "uuid": "c7383938-46cd-4a21-a97d-80b55d8b7888", 00:10:02.618 "assigned_rate_limits": { 00:10:02.618 "rw_ios_per_sec": 0, 00:10:02.618 "rw_mbytes_per_sec": 0, 00:10:02.618 "r_mbytes_per_sec": 0, 00:10:02.618 "w_mbytes_per_sec": 0 00:10:02.618 }, 00:10:02.618 "claimed": true, 00:10:02.618 "claim_type": "exclusive_write", 00:10:02.618 "zoned": false, 00:10:02.618 "supported_io_types": { 00:10:02.618 "read": true, 00:10:02.618 "write": true, 00:10:02.618 "unmap": true, 00:10:02.618 "flush": true, 00:10:02.618 "reset": true, 00:10:02.618 "nvme_admin": false, 00:10:02.618 "nvme_io": false, 00:10:02.618 "nvme_io_md": false, 00:10:02.618 "write_zeroes": true, 00:10:02.618 "zcopy": true, 00:10:02.618 "get_zone_info": false, 00:10:02.618 "zone_management": false, 00:10:02.618 "zone_append": false, 00:10:02.618 "compare": false, 00:10:02.618 "compare_and_write": false, 00:10:02.618 "abort": true, 00:10:02.618 "seek_hole": false, 00:10:02.618 "seek_data": false, 00:10:02.618 "copy": true, 00:10:02.618 "nvme_iov_md": false 00:10:02.618 }, 00:10:02.618 "memory_domains": [ 00:10:02.618 { 00:10:02.618 "dma_device_id": "system", 00:10:02.618 "dma_device_type": 1 00:10:02.618 }, 00:10:02.618 { 00:10:02.618 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:02.618 "dma_device_type": 2 00:10:02.618 } 00:10:02.618 ], 00:10:02.618 "driver_specific": {} 00:10:02.618 } 00:10:02.618 ] 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:02.618 "name": "Existed_Raid", 00:10:02.618 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:02.618 "strip_size_kb": 64, 00:10:02.618 "state": "configuring", 00:10:02.618 "raid_level": "concat", 00:10:02.618 "superblock": false, 00:10:02.618 "num_base_bdevs": 3, 00:10:02.618 "num_base_bdevs_discovered": 2, 00:10:02.618 "num_base_bdevs_operational": 3, 00:10:02.618 "base_bdevs_list": [ 00:10:02.618 { 00:10:02.618 "name": "BaseBdev1", 00:10:02.618 "uuid": "c7383938-46cd-4a21-a97d-80b55d8b7888", 00:10:02.618 "is_configured": true, 00:10:02.618 "data_offset": 0, 00:10:02.618 "data_size": 65536 00:10:02.618 }, 00:10:02.618 { 00:10:02.618 "name": null, 00:10:02.618 "uuid": "e7de92a6-90de-4e89-b52e-b3c949f8f0dd", 00:10:02.618 "is_configured": false, 00:10:02.618 "data_offset": 0, 00:10:02.618 "data_size": 65536 00:10:02.618 }, 00:10:02.618 { 00:10:02.618 "name": "BaseBdev3", 00:10:02.618 "uuid": "73885ed7-0d71-44d2-a3f9-239d19e04710", 00:10:02.618 "is_configured": true, 00:10:02.618 "data_offset": 0, 00:10:02.618 "data_size": 65536 00:10:02.618 } 00:10:02.618 ] 00:10:02.618 }' 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:02.618 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.876 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.876 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.876 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.876 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:02.876 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.876 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:02.876 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:02.876 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.876 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.876 [2024-11-20 15:58:00.973627] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:02.876 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.876 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:02.876 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:02.876 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:02.877 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:02.877 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:02.877 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:02.877 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:02.877 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:02.877 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:02.877 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:02.877 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.877 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.877 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.877 15:58:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:02.877 15:58:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.877 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:02.877 "name": "Existed_Raid", 00:10:02.877 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:02.877 "strip_size_kb": 64, 00:10:02.877 "state": "configuring", 00:10:02.877 "raid_level": "concat", 00:10:02.877 "superblock": false, 00:10:02.877 "num_base_bdevs": 3, 00:10:02.877 "num_base_bdevs_discovered": 1, 00:10:02.877 "num_base_bdevs_operational": 3, 00:10:02.877 "base_bdevs_list": [ 00:10:02.877 { 00:10:02.877 "name": "BaseBdev1", 00:10:02.877 "uuid": "c7383938-46cd-4a21-a97d-80b55d8b7888", 00:10:02.877 "is_configured": true, 00:10:02.877 "data_offset": 0, 00:10:02.877 "data_size": 65536 00:10:02.877 }, 00:10:02.877 { 00:10:02.877 "name": null, 00:10:02.877 "uuid": "e7de92a6-90de-4e89-b52e-b3c949f8f0dd", 00:10:02.877 "is_configured": false, 00:10:02.877 "data_offset": 0, 00:10:02.877 "data_size": 65536 00:10:02.877 }, 00:10:02.877 { 00:10:02.877 "name": null, 00:10:02.877 "uuid": "73885ed7-0d71-44d2-a3f9-239d19e04710", 00:10:02.877 "is_configured": false, 00:10:02.877 "data_offset": 0, 00:10:02.877 "data_size": 65536 00:10:02.877 } 00:10:02.877 ] 00:10:02.877 }' 00:10:02.877 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:02.877 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.135 [2024-11-20 15:58:01.285725] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:03.135 "name": "Existed_Raid", 00:10:03.135 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:03.135 "strip_size_kb": 64, 00:10:03.135 "state": "configuring", 00:10:03.135 "raid_level": "concat", 00:10:03.135 "superblock": false, 00:10:03.135 "num_base_bdevs": 3, 00:10:03.135 "num_base_bdevs_discovered": 2, 00:10:03.135 "num_base_bdevs_operational": 3, 00:10:03.135 "base_bdevs_list": [ 00:10:03.135 { 00:10:03.135 "name": "BaseBdev1", 00:10:03.135 "uuid": "c7383938-46cd-4a21-a97d-80b55d8b7888", 00:10:03.135 "is_configured": true, 00:10:03.135 "data_offset": 0, 00:10:03.135 "data_size": 65536 00:10:03.135 }, 00:10:03.135 { 00:10:03.135 "name": null, 00:10:03.135 "uuid": "e7de92a6-90de-4e89-b52e-b3c949f8f0dd", 00:10:03.135 "is_configured": false, 00:10:03.135 "data_offset": 0, 00:10:03.135 "data_size": 65536 00:10:03.135 }, 00:10:03.135 { 00:10:03.135 "name": "BaseBdev3", 00:10:03.135 "uuid": "73885ed7-0d71-44d2-a3f9-239d19e04710", 00:10:03.135 "is_configured": true, 00:10:03.135 "data_offset": 0, 00:10:03.135 "data_size": 65536 00:10:03.135 } 00:10:03.135 ] 00:10:03.135 }' 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:03.135 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.394 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:03.394 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.394 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.394 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.394 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.394 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:03.394 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:03.394 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.394 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.394 [2024-11-20 15:58:01.621781] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:03.651 "name": "Existed_Raid", 00:10:03.651 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:03.651 "strip_size_kb": 64, 00:10:03.651 "state": "configuring", 00:10:03.651 "raid_level": "concat", 00:10:03.651 "superblock": false, 00:10:03.651 "num_base_bdevs": 3, 00:10:03.651 "num_base_bdevs_discovered": 1, 00:10:03.651 "num_base_bdevs_operational": 3, 00:10:03.651 "base_bdevs_list": [ 00:10:03.651 { 00:10:03.651 "name": null, 00:10:03.651 "uuid": "c7383938-46cd-4a21-a97d-80b55d8b7888", 00:10:03.651 "is_configured": false, 00:10:03.651 "data_offset": 0, 00:10:03.651 "data_size": 65536 00:10:03.651 }, 00:10:03.651 { 00:10:03.651 "name": null, 00:10:03.651 "uuid": "e7de92a6-90de-4e89-b52e-b3c949f8f0dd", 00:10:03.651 "is_configured": false, 00:10:03.651 "data_offset": 0, 00:10:03.651 "data_size": 65536 00:10:03.651 }, 00:10:03.651 { 00:10:03.651 "name": "BaseBdev3", 00:10:03.651 "uuid": "73885ed7-0d71-44d2-a3f9-239d19e04710", 00:10:03.651 "is_configured": true, 00:10:03.651 "data_offset": 0, 00:10:03.651 "data_size": 65536 00:10:03.651 } 00:10:03.651 ] 00:10:03.651 }' 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:03.651 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.909 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.909 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.909 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.909 15:58:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:03.909 15:58:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.909 [2024-11-20 15:58:02.011003] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.909 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:03.909 "name": "Existed_Raid", 00:10:03.909 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:03.909 "strip_size_kb": 64, 00:10:03.909 "state": "configuring", 00:10:03.909 "raid_level": "concat", 00:10:03.909 "superblock": false, 00:10:03.909 "num_base_bdevs": 3, 00:10:03.910 "num_base_bdevs_discovered": 2, 00:10:03.910 "num_base_bdevs_operational": 3, 00:10:03.910 "base_bdevs_list": [ 00:10:03.910 { 00:10:03.910 "name": null, 00:10:03.910 "uuid": "c7383938-46cd-4a21-a97d-80b55d8b7888", 00:10:03.910 "is_configured": false, 00:10:03.910 "data_offset": 0, 00:10:03.910 "data_size": 65536 00:10:03.910 }, 00:10:03.910 { 00:10:03.910 "name": "BaseBdev2", 00:10:03.910 "uuid": "e7de92a6-90de-4e89-b52e-b3c949f8f0dd", 00:10:03.910 "is_configured": true, 00:10:03.910 "data_offset": 0, 00:10:03.910 "data_size": 65536 00:10:03.910 }, 00:10:03.910 { 00:10:03.910 "name": "BaseBdev3", 00:10:03.910 "uuid": "73885ed7-0d71-44d2-a3f9-239d19e04710", 00:10:03.910 "is_configured": true, 00:10:03.910 "data_offset": 0, 00:10:03.910 "data_size": 65536 00:10:03.910 } 00:10:03.910 ] 00:10:03.910 }' 00:10:03.910 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:03.910 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.167 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:04.167 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.167 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.167 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.167 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.167 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:04.167 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.167 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:04.167 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.167 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.167 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.167 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u c7383938-46cd-4a21-a97d-80b55d8b7888 00:10:04.167 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.167 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.167 [2024-11-20 15:58:02.389182] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:04.167 [2024-11-20 15:58:02.389210] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:10:04.167 [2024-11-20 15:58:02.389217] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:10:04.167 [2024-11-20 15:58:02.389411] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:10:04.167 [2024-11-20 15:58:02.389509] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:10:04.167 [2024-11-20 15:58:02.389515] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:10:04.167 [2024-11-20 15:58:02.389704] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:04.168 NewBaseBdev 00:10:04.168 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.168 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:04.168 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:04.168 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:04.168 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:04.168 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:04.168 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:04.168 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:04.168 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.168 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.168 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.168 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:04.168 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.168 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.426 [ 00:10:04.426 { 00:10:04.426 "name": "NewBaseBdev", 00:10:04.426 "aliases": [ 00:10:04.426 "c7383938-46cd-4a21-a97d-80b55d8b7888" 00:10:04.426 ], 00:10:04.426 "product_name": "Malloc disk", 00:10:04.426 "block_size": 512, 00:10:04.426 "num_blocks": 65536, 00:10:04.426 "uuid": "c7383938-46cd-4a21-a97d-80b55d8b7888", 00:10:04.426 "assigned_rate_limits": { 00:10:04.426 "rw_ios_per_sec": 0, 00:10:04.426 "rw_mbytes_per_sec": 0, 00:10:04.426 "r_mbytes_per_sec": 0, 00:10:04.426 "w_mbytes_per_sec": 0 00:10:04.426 }, 00:10:04.426 "claimed": true, 00:10:04.426 "claim_type": "exclusive_write", 00:10:04.426 "zoned": false, 00:10:04.426 "supported_io_types": { 00:10:04.426 "read": true, 00:10:04.426 "write": true, 00:10:04.426 "unmap": true, 00:10:04.426 "flush": true, 00:10:04.426 "reset": true, 00:10:04.426 "nvme_admin": false, 00:10:04.426 "nvme_io": false, 00:10:04.426 "nvme_io_md": false, 00:10:04.426 "write_zeroes": true, 00:10:04.426 "zcopy": true, 00:10:04.426 "get_zone_info": false, 00:10:04.426 "zone_management": false, 00:10:04.426 "zone_append": false, 00:10:04.426 "compare": false, 00:10:04.426 "compare_and_write": false, 00:10:04.426 "abort": true, 00:10:04.426 "seek_hole": false, 00:10:04.426 "seek_data": false, 00:10:04.426 "copy": true, 00:10:04.426 "nvme_iov_md": false 00:10:04.426 }, 00:10:04.426 "memory_domains": [ 00:10:04.426 { 00:10:04.426 "dma_device_id": "system", 00:10:04.426 "dma_device_type": 1 00:10:04.426 }, 00:10:04.426 { 00:10:04.426 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.426 "dma_device_type": 2 00:10:04.426 } 00:10:04.426 ], 00:10:04.426 "driver_specific": {} 00:10:04.426 } 00:10:04.426 ] 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:04.426 "name": "Existed_Raid", 00:10:04.426 "uuid": "3219d580-3ede-4dde-936e-727e481e6db0", 00:10:04.426 "strip_size_kb": 64, 00:10:04.426 "state": "online", 00:10:04.426 "raid_level": "concat", 00:10:04.426 "superblock": false, 00:10:04.426 "num_base_bdevs": 3, 00:10:04.426 "num_base_bdevs_discovered": 3, 00:10:04.426 "num_base_bdevs_operational": 3, 00:10:04.426 "base_bdevs_list": [ 00:10:04.426 { 00:10:04.426 "name": "NewBaseBdev", 00:10:04.426 "uuid": "c7383938-46cd-4a21-a97d-80b55d8b7888", 00:10:04.426 "is_configured": true, 00:10:04.426 "data_offset": 0, 00:10:04.426 "data_size": 65536 00:10:04.426 }, 00:10:04.426 { 00:10:04.426 "name": "BaseBdev2", 00:10:04.426 "uuid": "e7de92a6-90de-4e89-b52e-b3c949f8f0dd", 00:10:04.426 "is_configured": true, 00:10:04.426 "data_offset": 0, 00:10:04.426 "data_size": 65536 00:10:04.426 }, 00:10:04.426 { 00:10:04.426 "name": "BaseBdev3", 00:10:04.426 "uuid": "73885ed7-0d71-44d2-a3f9-239d19e04710", 00:10:04.426 "is_configured": true, 00:10:04.426 "data_offset": 0, 00:10:04.426 "data_size": 65536 00:10:04.426 } 00:10:04.426 ] 00:10:04.426 }' 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:04.426 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.683 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:04.683 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:04.683 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:04.683 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:04.683 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:04.683 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:04.683 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:04.683 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:04.683 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.683 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.683 [2024-11-20 15:58:02.737551] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:04.683 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.683 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:04.683 "name": "Existed_Raid", 00:10:04.683 "aliases": [ 00:10:04.683 "3219d580-3ede-4dde-936e-727e481e6db0" 00:10:04.683 ], 00:10:04.683 "product_name": "Raid Volume", 00:10:04.683 "block_size": 512, 00:10:04.683 "num_blocks": 196608, 00:10:04.683 "uuid": "3219d580-3ede-4dde-936e-727e481e6db0", 00:10:04.683 "assigned_rate_limits": { 00:10:04.683 "rw_ios_per_sec": 0, 00:10:04.683 "rw_mbytes_per_sec": 0, 00:10:04.683 "r_mbytes_per_sec": 0, 00:10:04.683 "w_mbytes_per_sec": 0 00:10:04.683 }, 00:10:04.683 "claimed": false, 00:10:04.683 "zoned": false, 00:10:04.683 "supported_io_types": { 00:10:04.683 "read": true, 00:10:04.683 "write": true, 00:10:04.684 "unmap": true, 00:10:04.684 "flush": true, 00:10:04.684 "reset": true, 00:10:04.684 "nvme_admin": false, 00:10:04.684 "nvme_io": false, 00:10:04.684 "nvme_io_md": false, 00:10:04.684 "write_zeroes": true, 00:10:04.684 "zcopy": false, 00:10:04.684 "get_zone_info": false, 00:10:04.684 "zone_management": false, 00:10:04.684 "zone_append": false, 00:10:04.684 "compare": false, 00:10:04.684 "compare_and_write": false, 00:10:04.684 "abort": false, 00:10:04.684 "seek_hole": false, 00:10:04.684 "seek_data": false, 00:10:04.684 "copy": false, 00:10:04.684 "nvme_iov_md": false 00:10:04.684 }, 00:10:04.684 "memory_domains": [ 00:10:04.684 { 00:10:04.684 "dma_device_id": "system", 00:10:04.684 "dma_device_type": 1 00:10:04.684 }, 00:10:04.684 { 00:10:04.684 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.684 "dma_device_type": 2 00:10:04.684 }, 00:10:04.684 { 00:10:04.684 "dma_device_id": "system", 00:10:04.684 "dma_device_type": 1 00:10:04.684 }, 00:10:04.684 { 00:10:04.684 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.684 "dma_device_type": 2 00:10:04.684 }, 00:10:04.684 { 00:10:04.684 "dma_device_id": "system", 00:10:04.684 "dma_device_type": 1 00:10:04.684 }, 00:10:04.684 { 00:10:04.684 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.684 "dma_device_type": 2 00:10:04.684 } 00:10:04.684 ], 00:10:04.684 "driver_specific": { 00:10:04.684 "raid": { 00:10:04.684 "uuid": "3219d580-3ede-4dde-936e-727e481e6db0", 00:10:04.684 "strip_size_kb": 64, 00:10:04.684 "state": "online", 00:10:04.684 "raid_level": "concat", 00:10:04.684 "superblock": false, 00:10:04.684 "num_base_bdevs": 3, 00:10:04.684 "num_base_bdevs_discovered": 3, 00:10:04.684 "num_base_bdevs_operational": 3, 00:10:04.684 "base_bdevs_list": [ 00:10:04.684 { 00:10:04.684 "name": "NewBaseBdev", 00:10:04.684 "uuid": "c7383938-46cd-4a21-a97d-80b55d8b7888", 00:10:04.684 "is_configured": true, 00:10:04.684 "data_offset": 0, 00:10:04.684 "data_size": 65536 00:10:04.684 }, 00:10:04.684 { 00:10:04.684 "name": "BaseBdev2", 00:10:04.684 "uuid": "e7de92a6-90de-4e89-b52e-b3c949f8f0dd", 00:10:04.684 "is_configured": true, 00:10:04.684 "data_offset": 0, 00:10:04.684 "data_size": 65536 00:10:04.684 }, 00:10:04.684 { 00:10:04.684 "name": "BaseBdev3", 00:10:04.684 "uuid": "73885ed7-0d71-44d2-a3f9-239d19e04710", 00:10:04.684 "is_configured": true, 00:10:04.684 "data_offset": 0, 00:10:04.684 "data_size": 65536 00:10:04.684 } 00:10:04.684 ] 00:10:04.684 } 00:10:04.684 } 00:10:04.684 }' 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:04.684 BaseBdev2 00:10:04.684 BaseBdev3' 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.684 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.942 [2024-11-20 15:58:02.933325] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:04.942 [2024-11-20 15:58:02.933345] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:04.942 [2024-11-20 15:58:02.933397] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:04.942 [2024-11-20 15:58:02.933445] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:04.942 [2024-11-20 15:58:02.933455] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:10:04.942 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.942 15:58:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 64018 00:10:04.942 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 64018 ']' 00:10:04.942 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 64018 00:10:04.942 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:10:04.942 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:04.942 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 64018 00:10:04.942 killing process with pid 64018 00:10:04.942 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:04.942 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:04.942 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 64018' 00:10:04.942 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 64018 00:10:04.942 [2024-11-20 15:58:02.965654] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:04.942 15:58:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 64018 00:10:04.942 [2024-11-20 15:58:03.111724] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:10:05.507 00:10:05.507 real 0m7.327s 00:10:05.507 user 0m11.807s 00:10:05.507 sys 0m1.219s 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:05.507 ************************************ 00:10:05.507 END TEST raid_state_function_test 00:10:05.507 ************************************ 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.507 15:58:03 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 3 true 00:10:05.507 15:58:03 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:05.507 15:58:03 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:05.507 15:58:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:05.507 ************************************ 00:10:05.507 START TEST raid_state_function_test_sb 00:10:05.507 ************************************ 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 3 true 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:05.507 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:10:05.508 Process raid pid: 64606 00:10:05.508 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=64606 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 64606' 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 64606 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 64606 ']' 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.508 15:58:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:05.765 [2024-11-20 15:58:03.787080] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:10:05.765 [2024-11-20 15:58:03.787171] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:05.765 [2024-11-20 15:58:03.937897] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:06.023 [2024-11-20 15:58:04.037444] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:06.023 [2024-11-20 15:58:04.178660] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:06.023 [2024-11-20 15:58:04.178695] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:06.588 15:58:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:06.588 15:58:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:10:06.588 15:58:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:06.588 15:58:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.588 15:58:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.588 [2024-11-20 15:58:04.661393] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:06.588 [2024-11-20 15:58:04.661445] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:06.588 [2024-11-20 15:58:04.661456] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:06.588 [2024-11-20 15:58:04.661465] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:06.588 [2024-11-20 15:58:04.661472] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:06.588 [2024-11-20 15:58:04.661480] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:06.588 15:58:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.588 15:58:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:06.589 15:58:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:06.589 15:58:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:06.589 15:58:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:06.589 15:58:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:06.589 15:58:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:06.589 15:58:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:06.589 15:58:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:06.589 15:58:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:06.589 15:58:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:06.589 15:58:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.589 15:58:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:06.589 15:58:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.589 15:58:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.589 15:58:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.589 15:58:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:06.589 "name": "Existed_Raid", 00:10:06.589 "uuid": "bef66133-71ce-4083-9d0f-60886c958163", 00:10:06.589 "strip_size_kb": 64, 00:10:06.589 "state": "configuring", 00:10:06.589 "raid_level": "concat", 00:10:06.589 "superblock": true, 00:10:06.589 "num_base_bdevs": 3, 00:10:06.589 "num_base_bdevs_discovered": 0, 00:10:06.589 "num_base_bdevs_operational": 3, 00:10:06.589 "base_bdevs_list": [ 00:10:06.589 { 00:10:06.589 "name": "BaseBdev1", 00:10:06.589 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:06.589 "is_configured": false, 00:10:06.589 "data_offset": 0, 00:10:06.589 "data_size": 0 00:10:06.589 }, 00:10:06.589 { 00:10:06.589 "name": "BaseBdev2", 00:10:06.589 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:06.589 "is_configured": false, 00:10:06.589 "data_offset": 0, 00:10:06.589 "data_size": 0 00:10:06.589 }, 00:10:06.589 { 00:10:06.589 "name": "BaseBdev3", 00:10:06.589 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:06.589 "is_configured": false, 00:10:06.589 "data_offset": 0, 00:10:06.589 "data_size": 0 00:10:06.589 } 00:10:06.589 ] 00:10:06.589 }' 00:10:06.589 15:58:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:06.589 15:58:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.847 15:58:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:06.847 15:58:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.847 15:58:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.847 [2024-11-20 15:58:04.993412] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:06.847 [2024-11-20 15:58:04.993445] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:10:06.847 15:58:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.847 15:58:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:06.847 15:58:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.847 15:58:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.847 [2024-11-20 15:58:05.001419] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:06.847 [2024-11-20 15:58:05.001461] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:06.847 [2024-11-20 15:58:05.001469] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:06.847 [2024-11-20 15:58:05.001479] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:06.847 [2024-11-20 15:58:05.001485] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:06.847 [2024-11-20 15:58:05.001493] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.847 [2024-11-20 15:58:05.033907] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:06.847 BaseBdev1 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.847 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.847 [ 00:10:06.848 { 00:10:06.848 "name": "BaseBdev1", 00:10:06.848 "aliases": [ 00:10:06.848 "862713f9-c7ab-4032-b06d-d748e5589200" 00:10:06.848 ], 00:10:06.848 "product_name": "Malloc disk", 00:10:06.848 "block_size": 512, 00:10:06.848 "num_blocks": 65536, 00:10:06.848 "uuid": "862713f9-c7ab-4032-b06d-d748e5589200", 00:10:06.848 "assigned_rate_limits": { 00:10:06.848 "rw_ios_per_sec": 0, 00:10:06.848 "rw_mbytes_per_sec": 0, 00:10:06.848 "r_mbytes_per_sec": 0, 00:10:06.848 "w_mbytes_per_sec": 0 00:10:06.848 }, 00:10:06.848 "claimed": true, 00:10:06.848 "claim_type": "exclusive_write", 00:10:06.848 "zoned": false, 00:10:06.848 "supported_io_types": { 00:10:06.848 "read": true, 00:10:06.848 "write": true, 00:10:06.848 "unmap": true, 00:10:06.848 "flush": true, 00:10:06.848 "reset": true, 00:10:06.848 "nvme_admin": false, 00:10:06.848 "nvme_io": false, 00:10:06.848 "nvme_io_md": false, 00:10:06.848 "write_zeroes": true, 00:10:06.848 "zcopy": true, 00:10:06.848 "get_zone_info": false, 00:10:06.848 "zone_management": false, 00:10:06.848 "zone_append": false, 00:10:06.848 "compare": false, 00:10:06.848 "compare_and_write": false, 00:10:06.848 "abort": true, 00:10:06.848 "seek_hole": false, 00:10:06.848 "seek_data": false, 00:10:06.848 "copy": true, 00:10:06.848 "nvme_iov_md": false 00:10:06.848 }, 00:10:06.848 "memory_domains": [ 00:10:06.848 { 00:10:06.848 "dma_device_id": "system", 00:10:06.848 "dma_device_type": 1 00:10:06.848 }, 00:10:06.848 { 00:10:06.848 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:06.848 "dma_device_type": 2 00:10:06.848 } 00:10:06.848 ], 00:10:06.848 "driver_specific": {} 00:10:06.848 } 00:10:06.848 ] 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:06.848 "name": "Existed_Raid", 00:10:06.848 "uuid": "65e3b848-f129-4f3f-912e-6fbd3da836ce", 00:10:06.848 "strip_size_kb": 64, 00:10:06.848 "state": "configuring", 00:10:06.848 "raid_level": "concat", 00:10:06.848 "superblock": true, 00:10:06.848 "num_base_bdevs": 3, 00:10:06.848 "num_base_bdevs_discovered": 1, 00:10:06.848 "num_base_bdevs_operational": 3, 00:10:06.848 "base_bdevs_list": [ 00:10:06.848 { 00:10:06.848 "name": "BaseBdev1", 00:10:06.848 "uuid": "862713f9-c7ab-4032-b06d-d748e5589200", 00:10:06.848 "is_configured": true, 00:10:06.848 "data_offset": 2048, 00:10:06.848 "data_size": 63488 00:10:06.848 }, 00:10:06.848 { 00:10:06.848 "name": "BaseBdev2", 00:10:06.848 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:06.848 "is_configured": false, 00:10:06.848 "data_offset": 0, 00:10:06.848 "data_size": 0 00:10:06.848 }, 00:10:06.848 { 00:10:06.848 "name": "BaseBdev3", 00:10:06.848 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:06.848 "is_configured": false, 00:10:06.848 "data_offset": 0, 00:10:06.848 "data_size": 0 00:10:06.848 } 00:10:06.848 ] 00:10:06.848 }' 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:06.848 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.227 [2024-11-20 15:58:05.362036] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:07.227 [2024-11-20 15:58:05.362199] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.227 [2024-11-20 15:58:05.370089] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:07.227 [2024-11-20 15:58:05.371919] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:07.227 [2024-11-20 15:58:05.372050] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:07.227 [2024-11-20 15:58:05.372065] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:07.227 [2024-11-20 15:58:05.372075] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:07.227 "name": "Existed_Raid", 00:10:07.227 "uuid": "c751e4aa-7d77-4ce8-8f49-47a807d113fe", 00:10:07.227 "strip_size_kb": 64, 00:10:07.227 "state": "configuring", 00:10:07.227 "raid_level": "concat", 00:10:07.227 "superblock": true, 00:10:07.227 "num_base_bdevs": 3, 00:10:07.227 "num_base_bdevs_discovered": 1, 00:10:07.227 "num_base_bdevs_operational": 3, 00:10:07.227 "base_bdevs_list": [ 00:10:07.227 { 00:10:07.227 "name": "BaseBdev1", 00:10:07.227 "uuid": "862713f9-c7ab-4032-b06d-d748e5589200", 00:10:07.227 "is_configured": true, 00:10:07.227 "data_offset": 2048, 00:10:07.227 "data_size": 63488 00:10:07.227 }, 00:10:07.227 { 00:10:07.227 "name": "BaseBdev2", 00:10:07.227 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:07.227 "is_configured": false, 00:10:07.227 "data_offset": 0, 00:10:07.227 "data_size": 0 00:10:07.227 }, 00:10:07.227 { 00:10:07.227 "name": "BaseBdev3", 00:10:07.227 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:07.227 "is_configured": false, 00:10:07.227 "data_offset": 0, 00:10:07.227 "data_size": 0 00:10:07.227 } 00:10:07.227 ] 00:10:07.227 }' 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:07.227 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.486 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:07.486 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.486 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.486 [2024-11-20 15:58:05.716283] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:07.486 BaseBdev2 00:10:07.486 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.486 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:07.486 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:07.486 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:07.486 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:07.486 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:07.486 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:07.486 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:07.486 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.486 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.486 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.486 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:07.486 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.486 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.744 [ 00:10:07.744 { 00:10:07.744 "name": "BaseBdev2", 00:10:07.744 "aliases": [ 00:10:07.744 "7f25bd52-4410-4f4e-a95a-dca131b4dc1f" 00:10:07.744 ], 00:10:07.744 "product_name": "Malloc disk", 00:10:07.744 "block_size": 512, 00:10:07.744 "num_blocks": 65536, 00:10:07.745 "uuid": "7f25bd52-4410-4f4e-a95a-dca131b4dc1f", 00:10:07.745 "assigned_rate_limits": { 00:10:07.745 "rw_ios_per_sec": 0, 00:10:07.745 "rw_mbytes_per_sec": 0, 00:10:07.745 "r_mbytes_per_sec": 0, 00:10:07.745 "w_mbytes_per_sec": 0 00:10:07.745 }, 00:10:07.745 "claimed": true, 00:10:07.745 "claim_type": "exclusive_write", 00:10:07.745 "zoned": false, 00:10:07.745 "supported_io_types": { 00:10:07.745 "read": true, 00:10:07.745 "write": true, 00:10:07.745 "unmap": true, 00:10:07.745 "flush": true, 00:10:07.745 "reset": true, 00:10:07.745 "nvme_admin": false, 00:10:07.745 "nvme_io": false, 00:10:07.745 "nvme_io_md": false, 00:10:07.745 "write_zeroes": true, 00:10:07.745 "zcopy": true, 00:10:07.745 "get_zone_info": false, 00:10:07.745 "zone_management": false, 00:10:07.745 "zone_append": false, 00:10:07.745 "compare": false, 00:10:07.745 "compare_and_write": false, 00:10:07.745 "abort": true, 00:10:07.745 "seek_hole": false, 00:10:07.745 "seek_data": false, 00:10:07.745 "copy": true, 00:10:07.745 "nvme_iov_md": false 00:10:07.745 }, 00:10:07.745 "memory_domains": [ 00:10:07.745 { 00:10:07.745 "dma_device_id": "system", 00:10:07.745 "dma_device_type": 1 00:10:07.745 }, 00:10:07.745 { 00:10:07.745 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:07.745 "dma_device_type": 2 00:10:07.745 } 00:10:07.745 ], 00:10:07.745 "driver_specific": {} 00:10:07.745 } 00:10:07.745 ] 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:07.745 "name": "Existed_Raid", 00:10:07.745 "uuid": "c751e4aa-7d77-4ce8-8f49-47a807d113fe", 00:10:07.745 "strip_size_kb": 64, 00:10:07.745 "state": "configuring", 00:10:07.745 "raid_level": "concat", 00:10:07.745 "superblock": true, 00:10:07.745 "num_base_bdevs": 3, 00:10:07.745 "num_base_bdevs_discovered": 2, 00:10:07.745 "num_base_bdevs_operational": 3, 00:10:07.745 "base_bdevs_list": [ 00:10:07.745 { 00:10:07.745 "name": "BaseBdev1", 00:10:07.745 "uuid": "862713f9-c7ab-4032-b06d-d748e5589200", 00:10:07.745 "is_configured": true, 00:10:07.745 "data_offset": 2048, 00:10:07.745 "data_size": 63488 00:10:07.745 }, 00:10:07.745 { 00:10:07.745 "name": "BaseBdev2", 00:10:07.745 "uuid": "7f25bd52-4410-4f4e-a95a-dca131b4dc1f", 00:10:07.745 "is_configured": true, 00:10:07.745 "data_offset": 2048, 00:10:07.745 "data_size": 63488 00:10:07.745 }, 00:10:07.745 { 00:10:07.745 "name": "BaseBdev3", 00:10:07.745 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:07.745 "is_configured": false, 00:10:07.745 "data_offset": 0, 00:10:07.745 "data_size": 0 00:10:07.745 } 00:10:07.745 ] 00:10:07.745 }' 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:07.745 15:58:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.003 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:08.003 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.003 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.003 [2024-11-20 15:58:06.084428] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:08.003 [2024-11-20 15:58:06.084793] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:10:08.003 [2024-11-20 15:58:06.084948] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:10:08.003 BaseBdev3 00:10:08.003 [2024-11-20 15:58:06.085241] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:10:08.003 [2024-11-20 15:58:06.085383] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:10:08.003 [2024-11-20 15:58:06.085393] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:10:08.003 [2024-11-20 15:58:06.085516] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:08.003 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.003 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:08.003 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:08.003 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:08.003 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:08.003 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:08.003 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:08.003 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:08.003 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.003 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.003 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.003 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:08.003 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.003 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.003 [ 00:10:08.003 { 00:10:08.003 "name": "BaseBdev3", 00:10:08.003 "aliases": [ 00:10:08.003 "e4c27933-bdc2-4d39-9785-85816f15abf3" 00:10:08.003 ], 00:10:08.003 "product_name": "Malloc disk", 00:10:08.003 "block_size": 512, 00:10:08.003 "num_blocks": 65536, 00:10:08.003 "uuid": "e4c27933-bdc2-4d39-9785-85816f15abf3", 00:10:08.003 "assigned_rate_limits": { 00:10:08.003 "rw_ios_per_sec": 0, 00:10:08.003 "rw_mbytes_per_sec": 0, 00:10:08.003 "r_mbytes_per_sec": 0, 00:10:08.003 "w_mbytes_per_sec": 0 00:10:08.003 }, 00:10:08.003 "claimed": true, 00:10:08.003 "claim_type": "exclusive_write", 00:10:08.003 "zoned": false, 00:10:08.003 "supported_io_types": { 00:10:08.003 "read": true, 00:10:08.003 "write": true, 00:10:08.003 "unmap": true, 00:10:08.003 "flush": true, 00:10:08.003 "reset": true, 00:10:08.003 "nvme_admin": false, 00:10:08.003 "nvme_io": false, 00:10:08.003 "nvme_io_md": false, 00:10:08.003 "write_zeroes": true, 00:10:08.003 "zcopy": true, 00:10:08.003 "get_zone_info": false, 00:10:08.003 "zone_management": false, 00:10:08.003 "zone_append": false, 00:10:08.003 "compare": false, 00:10:08.003 "compare_and_write": false, 00:10:08.003 "abort": true, 00:10:08.003 "seek_hole": false, 00:10:08.003 "seek_data": false, 00:10:08.003 "copy": true, 00:10:08.003 "nvme_iov_md": false 00:10:08.003 }, 00:10:08.003 "memory_domains": [ 00:10:08.003 { 00:10:08.003 "dma_device_id": "system", 00:10:08.003 "dma_device_type": 1 00:10:08.003 }, 00:10:08.003 { 00:10:08.003 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:08.003 "dma_device_type": 2 00:10:08.003 } 00:10:08.003 ], 00:10:08.003 "driver_specific": {} 00:10:08.004 } 00:10:08.004 ] 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:08.004 "name": "Existed_Raid", 00:10:08.004 "uuid": "c751e4aa-7d77-4ce8-8f49-47a807d113fe", 00:10:08.004 "strip_size_kb": 64, 00:10:08.004 "state": "online", 00:10:08.004 "raid_level": "concat", 00:10:08.004 "superblock": true, 00:10:08.004 "num_base_bdevs": 3, 00:10:08.004 "num_base_bdevs_discovered": 3, 00:10:08.004 "num_base_bdevs_operational": 3, 00:10:08.004 "base_bdevs_list": [ 00:10:08.004 { 00:10:08.004 "name": "BaseBdev1", 00:10:08.004 "uuid": "862713f9-c7ab-4032-b06d-d748e5589200", 00:10:08.004 "is_configured": true, 00:10:08.004 "data_offset": 2048, 00:10:08.004 "data_size": 63488 00:10:08.004 }, 00:10:08.004 { 00:10:08.004 "name": "BaseBdev2", 00:10:08.004 "uuid": "7f25bd52-4410-4f4e-a95a-dca131b4dc1f", 00:10:08.004 "is_configured": true, 00:10:08.004 "data_offset": 2048, 00:10:08.004 "data_size": 63488 00:10:08.004 }, 00:10:08.004 { 00:10:08.004 "name": "BaseBdev3", 00:10:08.004 "uuid": "e4c27933-bdc2-4d39-9785-85816f15abf3", 00:10:08.004 "is_configured": true, 00:10:08.004 "data_offset": 2048, 00:10:08.004 "data_size": 63488 00:10:08.004 } 00:10:08.004 ] 00:10:08.004 }' 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:08.004 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.261 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:08.261 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:08.261 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:08.261 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:08.261 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:08.261 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:08.261 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:08.261 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.261 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.261 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:08.261 [2024-11-20 15:58:06.404872] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:08.261 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.261 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:08.261 "name": "Existed_Raid", 00:10:08.261 "aliases": [ 00:10:08.261 "c751e4aa-7d77-4ce8-8f49-47a807d113fe" 00:10:08.261 ], 00:10:08.261 "product_name": "Raid Volume", 00:10:08.261 "block_size": 512, 00:10:08.261 "num_blocks": 190464, 00:10:08.261 "uuid": "c751e4aa-7d77-4ce8-8f49-47a807d113fe", 00:10:08.261 "assigned_rate_limits": { 00:10:08.261 "rw_ios_per_sec": 0, 00:10:08.261 "rw_mbytes_per_sec": 0, 00:10:08.261 "r_mbytes_per_sec": 0, 00:10:08.261 "w_mbytes_per_sec": 0 00:10:08.261 }, 00:10:08.261 "claimed": false, 00:10:08.261 "zoned": false, 00:10:08.261 "supported_io_types": { 00:10:08.261 "read": true, 00:10:08.262 "write": true, 00:10:08.262 "unmap": true, 00:10:08.262 "flush": true, 00:10:08.262 "reset": true, 00:10:08.262 "nvme_admin": false, 00:10:08.262 "nvme_io": false, 00:10:08.262 "nvme_io_md": false, 00:10:08.262 "write_zeroes": true, 00:10:08.262 "zcopy": false, 00:10:08.262 "get_zone_info": false, 00:10:08.262 "zone_management": false, 00:10:08.262 "zone_append": false, 00:10:08.262 "compare": false, 00:10:08.262 "compare_and_write": false, 00:10:08.262 "abort": false, 00:10:08.262 "seek_hole": false, 00:10:08.262 "seek_data": false, 00:10:08.262 "copy": false, 00:10:08.262 "nvme_iov_md": false 00:10:08.262 }, 00:10:08.262 "memory_domains": [ 00:10:08.262 { 00:10:08.262 "dma_device_id": "system", 00:10:08.262 "dma_device_type": 1 00:10:08.262 }, 00:10:08.262 { 00:10:08.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:08.262 "dma_device_type": 2 00:10:08.262 }, 00:10:08.262 { 00:10:08.262 "dma_device_id": "system", 00:10:08.262 "dma_device_type": 1 00:10:08.262 }, 00:10:08.262 { 00:10:08.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:08.262 "dma_device_type": 2 00:10:08.262 }, 00:10:08.262 { 00:10:08.262 "dma_device_id": "system", 00:10:08.262 "dma_device_type": 1 00:10:08.262 }, 00:10:08.262 { 00:10:08.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:08.262 "dma_device_type": 2 00:10:08.262 } 00:10:08.262 ], 00:10:08.262 "driver_specific": { 00:10:08.262 "raid": { 00:10:08.262 "uuid": "c751e4aa-7d77-4ce8-8f49-47a807d113fe", 00:10:08.262 "strip_size_kb": 64, 00:10:08.262 "state": "online", 00:10:08.262 "raid_level": "concat", 00:10:08.262 "superblock": true, 00:10:08.262 "num_base_bdevs": 3, 00:10:08.262 "num_base_bdevs_discovered": 3, 00:10:08.262 "num_base_bdevs_operational": 3, 00:10:08.262 "base_bdevs_list": [ 00:10:08.262 { 00:10:08.262 "name": "BaseBdev1", 00:10:08.262 "uuid": "862713f9-c7ab-4032-b06d-d748e5589200", 00:10:08.262 "is_configured": true, 00:10:08.262 "data_offset": 2048, 00:10:08.262 "data_size": 63488 00:10:08.262 }, 00:10:08.262 { 00:10:08.262 "name": "BaseBdev2", 00:10:08.262 "uuid": "7f25bd52-4410-4f4e-a95a-dca131b4dc1f", 00:10:08.262 "is_configured": true, 00:10:08.262 "data_offset": 2048, 00:10:08.262 "data_size": 63488 00:10:08.262 }, 00:10:08.262 { 00:10:08.262 "name": "BaseBdev3", 00:10:08.262 "uuid": "e4c27933-bdc2-4d39-9785-85816f15abf3", 00:10:08.262 "is_configured": true, 00:10:08.262 "data_offset": 2048, 00:10:08.262 "data_size": 63488 00:10:08.262 } 00:10:08.262 ] 00:10:08.262 } 00:10:08.262 } 00:10:08.262 }' 00:10:08.262 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:08.262 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:08.262 BaseBdev2 00:10:08.262 BaseBdev3' 00:10:08.262 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:08.262 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:08.262 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:08.262 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:08.262 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:08.262 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.262 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.262 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.519 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:08.519 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:08.519 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:08.519 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:08.519 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.519 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:08.519 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.519 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.519 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:08.519 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:08.519 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:08.519 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:08.519 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.520 [2024-11-20 15:58:06.584608] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:08.520 [2024-11-20 15:58:06.584746] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:08.520 [2024-11-20 15:58:06.584805] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:08.520 "name": "Existed_Raid", 00:10:08.520 "uuid": "c751e4aa-7d77-4ce8-8f49-47a807d113fe", 00:10:08.520 "strip_size_kb": 64, 00:10:08.520 "state": "offline", 00:10:08.520 "raid_level": "concat", 00:10:08.520 "superblock": true, 00:10:08.520 "num_base_bdevs": 3, 00:10:08.520 "num_base_bdevs_discovered": 2, 00:10:08.520 "num_base_bdevs_operational": 2, 00:10:08.520 "base_bdevs_list": [ 00:10:08.520 { 00:10:08.520 "name": null, 00:10:08.520 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:08.520 "is_configured": false, 00:10:08.520 "data_offset": 0, 00:10:08.520 "data_size": 63488 00:10:08.520 }, 00:10:08.520 { 00:10:08.520 "name": "BaseBdev2", 00:10:08.520 "uuid": "7f25bd52-4410-4f4e-a95a-dca131b4dc1f", 00:10:08.520 "is_configured": true, 00:10:08.520 "data_offset": 2048, 00:10:08.520 "data_size": 63488 00:10:08.520 }, 00:10:08.520 { 00:10:08.520 "name": "BaseBdev3", 00:10:08.520 "uuid": "e4c27933-bdc2-4d39-9785-85816f15abf3", 00:10:08.520 "is_configured": true, 00:10:08.520 "data_offset": 2048, 00:10:08.520 "data_size": 63488 00:10:08.520 } 00:10:08.520 ] 00:10:08.520 }' 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:08.520 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.777 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:08.777 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:08.777 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:08.777 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:08.777 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.777 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.777 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.777 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:08.778 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:08.778 15:58:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:08.778 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.778 15:58:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.778 [2024-11-20 15:58:06.955131] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:08.778 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.778 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:08.778 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:08.778 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:08.778 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.778 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:08.778 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.036 [2024-11-20 15:58:07.053286] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:09.036 [2024-11-20 15:58:07.053325] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.036 BaseBdev2 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.036 [ 00:10:09.036 { 00:10:09.036 "name": "BaseBdev2", 00:10:09.036 "aliases": [ 00:10:09.036 "6ff47516-0270-4bd2-845d-fd41575caf37" 00:10:09.036 ], 00:10:09.036 "product_name": "Malloc disk", 00:10:09.036 "block_size": 512, 00:10:09.036 "num_blocks": 65536, 00:10:09.036 "uuid": "6ff47516-0270-4bd2-845d-fd41575caf37", 00:10:09.036 "assigned_rate_limits": { 00:10:09.036 "rw_ios_per_sec": 0, 00:10:09.036 "rw_mbytes_per_sec": 0, 00:10:09.036 "r_mbytes_per_sec": 0, 00:10:09.036 "w_mbytes_per_sec": 0 00:10:09.036 }, 00:10:09.036 "claimed": false, 00:10:09.036 "zoned": false, 00:10:09.036 "supported_io_types": { 00:10:09.036 "read": true, 00:10:09.036 "write": true, 00:10:09.036 "unmap": true, 00:10:09.036 "flush": true, 00:10:09.036 "reset": true, 00:10:09.036 "nvme_admin": false, 00:10:09.036 "nvme_io": false, 00:10:09.036 "nvme_io_md": false, 00:10:09.036 "write_zeroes": true, 00:10:09.036 "zcopy": true, 00:10:09.036 "get_zone_info": false, 00:10:09.036 "zone_management": false, 00:10:09.036 "zone_append": false, 00:10:09.036 "compare": false, 00:10:09.036 "compare_and_write": false, 00:10:09.036 "abort": true, 00:10:09.036 "seek_hole": false, 00:10:09.036 "seek_data": false, 00:10:09.036 "copy": true, 00:10:09.036 "nvme_iov_md": false 00:10:09.036 }, 00:10:09.036 "memory_domains": [ 00:10:09.036 { 00:10:09.036 "dma_device_id": "system", 00:10:09.036 "dma_device_type": 1 00:10:09.036 }, 00:10:09.036 { 00:10:09.036 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.036 "dma_device_type": 2 00:10:09.036 } 00:10:09.036 ], 00:10:09.036 "driver_specific": {} 00:10:09.036 } 00:10:09.036 ] 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.036 BaseBdev3 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.036 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.036 [ 00:10:09.036 { 00:10:09.036 "name": "BaseBdev3", 00:10:09.036 "aliases": [ 00:10:09.036 "fe55b125-51ec-4d2e-a7b8-981e37517c5c" 00:10:09.036 ], 00:10:09.036 "product_name": "Malloc disk", 00:10:09.036 "block_size": 512, 00:10:09.036 "num_blocks": 65536, 00:10:09.036 "uuid": "fe55b125-51ec-4d2e-a7b8-981e37517c5c", 00:10:09.036 "assigned_rate_limits": { 00:10:09.036 "rw_ios_per_sec": 0, 00:10:09.036 "rw_mbytes_per_sec": 0, 00:10:09.036 "r_mbytes_per_sec": 0, 00:10:09.036 "w_mbytes_per_sec": 0 00:10:09.036 }, 00:10:09.036 "claimed": false, 00:10:09.036 "zoned": false, 00:10:09.036 "supported_io_types": { 00:10:09.036 "read": true, 00:10:09.036 "write": true, 00:10:09.036 "unmap": true, 00:10:09.036 "flush": true, 00:10:09.036 "reset": true, 00:10:09.036 "nvme_admin": false, 00:10:09.036 "nvme_io": false, 00:10:09.036 "nvme_io_md": false, 00:10:09.036 "write_zeroes": true, 00:10:09.036 "zcopy": true, 00:10:09.036 "get_zone_info": false, 00:10:09.036 "zone_management": false, 00:10:09.036 "zone_append": false, 00:10:09.036 "compare": false, 00:10:09.036 "compare_and_write": false, 00:10:09.036 "abort": true, 00:10:09.036 "seek_hole": false, 00:10:09.036 "seek_data": false, 00:10:09.036 "copy": true, 00:10:09.037 "nvme_iov_md": false 00:10:09.037 }, 00:10:09.037 "memory_domains": [ 00:10:09.037 { 00:10:09.037 "dma_device_id": "system", 00:10:09.037 "dma_device_type": 1 00:10:09.037 }, 00:10:09.037 { 00:10:09.037 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.037 "dma_device_type": 2 00:10:09.037 } 00:10:09.037 ], 00:10:09.037 "driver_specific": {} 00:10:09.037 } 00:10:09.037 ] 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.037 [2024-11-20 15:58:07.246055] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:09.037 [2024-11-20 15:58:07.246181] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:09.037 [2024-11-20 15:58:07.246250] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:09.037 [2024-11-20 15:58:07.248044] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:09.037 "name": "Existed_Raid", 00:10:09.037 "uuid": "b0a75242-ace3-4c3c-af0f-79a53721a07a", 00:10:09.037 "strip_size_kb": 64, 00:10:09.037 "state": "configuring", 00:10:09.037 "raid_level": "concat", 00:10:09.037 "superblock": true, 00:10:09.037 "num_base_bdevs": 3, 00:10:09.037 "num_base_bdevs_discovered": 2, 00:10:09.037 "num_base_bdevs_operational": 3, 00:10:09.037 "base_bdevs_list": [ 00:10:09.037 { 00:10:09.037 "name": "BaseBdev1", 00:10:09.037 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.037 "is_configured": false, 00:10:09.037 "data_offset": 0, 00:10:09.037 "data_size": 0 00:10:09.037 }, 00:10:09.037 { 00:10:09.037 "name": "BaseBdev2", 00:10:09.037 "uuid": "6ff47516-0270-4bd2-845d-fd41575caf37", 00:10:09.037 "is_configured": true, 00:10:09.037 "data_offset": 2048, 00:10:09.037 "data_size": 63488 00:10:09.037 }, 00:10:09.037 { 00:10:09.037 "name": "BaseBdev3", 00:10:09.037 "uuid": "fe55b125-51ec-4d2e-a7b8-981e37517c5c", 00:10:09.037 "is_configured": true, 00:10:09.037 "data_offset": 2048, 00:10:09.037 "data_size": 63488 00:10:09.037 } 00:10:09.037 ] 00:10:09.037 }' 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:09.037 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.602 [2024-11-20 15:58:07.550107] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:09.602 "name": "Existed_Raid", 00:10:09.602 "uuid": "b0a75242-ace3-4c3c-af0f-79a53721a07a", 00:10:09.602 "strip_size_kb": 64, 00:10:09.602 "state": "configuring", 00:10:09.602 "raid_level": "concat", 00:10:09.602 "superblock": true, 00:10:09.602 "num_base_bdevs": 3, 00:10:09.602 "num_base_bdevs_discovered": 1, 00:10:09.602 "num_base_bdevs_operational": 3, 00:10:09.602 "base_bdevs_list": [ 00:10:09.602 { 00:10:09.602 "name": "BaseBdev1", 00:10:09.602 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.602 "is_configured": false, 00:10:09.602 "data_offset": 0, 00:10:09.602 "data_size": 0 00:10:09.602 }, 00:10:09.602 { 00:10:09.602 "name": null, 00:10:09.602 "uuid": "6ff47516-0270-4bd2-845d-fd41575caf37", 00:10:09.602 "is_configured": false, 00:10:09.602 "data_offset": 0, 00:10:09.602 "data_size": 63488 00:10:09.602 }, 00:10:09.602 { 00:10:09.602 "name": "BaseBdev3", 00:10:09.602 "uuid": "fe55b125-51ec-4d2e-a7b8-981e37517c5c", 00:10:09.602 "is_configured": true, 00:10:09.602 "data_offset": 2048, 00:10:09.602 "data_size": 63488 00:10:09.602 } 00:10:09.602 ] 00:10:09.602 }' 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:09.602 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.860 [2024-11-20 15:58:07.912279] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:09.860 BaseBdev1 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.860 [ 00:10:09.860 { 00:10:09.860 "name": "BaseBdev1", 00:10:09.860 "aliases": [ 00:10:09.860 "7a28cce6-a01f-4951-bac0-f1e60acf918a" 00:10:09.860 ], 00:10:09.860 "product_name": "Malloc disk", 00:10:09.860 "block_size": 512, 00:10:09.860 "num_blocks": 65536, 00:10:09.860 "uuid": "7a28cce6-a01f-4951-bac0-f1e60acf918a", 00:10:09.860 "assigned_rate_limits": { 00:10:09.860 "rw_ios_per_sec": 0, 00:10:09.860 "rw_mbytes_per_sec": 0, 00:10:09.860 "r_mbytes_per_sec": 0, 00:10:09.860 "w_mbytes_per_sec": 0 00:10:09.860 }, 00:10:09.860 "claimed": true, 00:10:09.860 "claim_type": "exclusive_write", 00:10:09.860 "zoned": false, 00:10:09.860 "supported_io_types": { 00:10:09.860 "read": true, 00:10:09.860 "write": true, 00:10:09.860 "unmap": true, 00:10:09.860 "flush": true, 00:10:09.860 "reset": true, 00:10:09.860 "nvme_admin": false, 00:10:09.860 "nvme_io": false, 00:10:09.860 "nvme_io_md": false, 00:10:09.860 "write_zeroes": true, 00:10:09.860 "zcopy": true, 00:10:09.860 "get_zone_info": false, 00:10:09.860 "zone_management": false, 00:10:09.860 "zone_append": false, 00:10:09.860 "compare": false, 00:10:09.860 "compare_and_write": false, 00:10:09.860 "abort": true, 00:10:09.860 "seek_hole": false, 00:10:09.860 "seek_data": false, 00:10:09.860 "copy": true, 00:10:09.860 "nvme_iov_md": false 00:10:09.860 }, 00:10:09.860 "memory_domains": [ 00:10:09.860 { 00:10:09.860 "dma_device_id": "system", 00:10:09.860 "dma_device_type": 1 00:10:09.860 }, 00:10:09.860 { 00:10:09.860 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.860 "dma_device_type": 2 00:10:09.860 } 00:10:09.860 ], 00:10:09.860 "driver_specific": {} 00:10:09.860 } 00:10:09.860 ] 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.860 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:09.860 "name": "Existed_Raid", 00:10:09.860 "uuid": "b0a75242-ace3-4c3c-af0f-79a53721a07a", 00:10:09.860 "strip_size_kb": 64, 00:10:09.860 "state": "configuring", 00:10:09.860 "raid_level": "concat", 00:10:09.860 "superblock": true, 00:10:09.860 "num_base_bdevs": 3, 00:10:09.860 "num_base_bdevs_discovered": 2, 00:10:09.860 "num_base_bdevs_operational": 3, 00:10:09.860 "base_bdevs_list": [ 00:10:09.860 { 00:10:09.860 "name": "BaseBdev1", 00:10:09.860 "uuid": "7a28cce6-a01f-4951-bac0-f1e60acf918a", 00:10:09.860 "is_configured": true, 00:10:09.861 "data_offset": 2048, 00:10:09.861 "data_size": 63488 00:10:09.861 }, 00:10:09.861 { 00:10:09.861 "name": null, 00:10:09.861 "uuid": "6ff47516-0270-4bd2-845d-fd41575caf37", 00:10:09.861 "is_configured": false, 00:10:09.861 "data_offset": 0, 00:10:09.861 "data_size": 63488 00:10:09.861 }, 00:10:09.861 { 00:10:09.861 "name": "BaseBdev3", 00:10:09.861 "uuid": "fe55b125-51ec-4d2e-a7b8-981e37517c5c", 00:10:09.861 "is_configured": true, 00:10:09.861 "data_offset": 2048, 00:10:09.861 "data_size": 63488 00:10:09.861 } 00:10:09.861 ] 00:10:09.861 }' 00:10:09.861 15:58:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:09.861 15:58:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.118 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.118 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.118 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:10.118 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.118 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.118 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:10.118 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:10.118 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.118 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.118 [2024-11-20 15:58:08.284386] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:10.118 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.118 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:10.118 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:10.118 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:10.119 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:10.119 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:10.119 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:10.119 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:10.119 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:10.119 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:10.119 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:10.119 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.119 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.119 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.119 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:10.119 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.119 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:10.119 "name": "Existed_Raid", 00:10:10.119 "uuid": "b0a75242-ace3-4c3c-af0f-79a53721a07a", 00:10:10.119 "strip_size_kb": 64, 00:10:10.119 "state": "configuring", 00:10:10.119 "raid_level": "concat", 00:10:10.119 "superblock": true, 00:10:10.119 "num_base_bdevs": 3, 00:10:10.119 "num_base_bdevs_discovered": 1, 00:10:10.119 "num_base_bdevs_operational": 3, 00:10:10.119 "base_bdevs_list": [ 00:10:10.119 { 00:10:10.119 "name": "BaseBdev1", 00:10:10.119 "uuid": "7a28cce6-a01f-4951-bac0-f1e60acf918a", 00:10:10.119 "is_configured": true, 00:10:10.119 "data_offset": 2048, 00:10:10.119 "data_size": 63488 00:10:10.119 }, 00:10:10.119 { 00:10:10.119 "name": null, 00:10:10.119 "uuid": "6ff47516-0270-4bd2-845d-fd41575caf37", 00:10:10.119 "is_configured": false, 00:10:10.119 "data_offset": 0, 00:10:10.119 "data_size": 63488 00:10:10.119 }, 00:10:10.119 { 00:10:10.119 "name": null, 00:10:10.119 "uuid": "fe55b125-51ec-4d2e-a7b8-981e37517c5c", 00:10:10.119 "is_configured": false, 00:10:10.119 "data_offset": 0, 00:10:10.119 "data_size": 63488 00:10:10.119 } 00:10:10.119 ] 00:10:10.119 }' 00:10:10.119 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:10.119 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.376 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.376 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:10.376 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.376 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.376 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.376 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:10.376 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:10.376 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.376 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.376 [2024-11-20 15:58:08.624474] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:10.634 "name": "Existed_Raid", 00:10:10.634 "uuid": "b0a75242-ace3-4c3c-af0f-79a53721a07a", 00:10:10.634 "strip_size_kb": 64, 00:10:10.634 "state": "configuring", 00:10:10.634 "raid_level": "concat", 00:10:10.634 "superblock": true, 00:10:10.634 "num_base_bdevs": 3, 00:10:10.634 "num_base_bdevs_discovered": 2, 00:10:10.634 "num_base_bdevs_operational": 3, 00:10:10.634 "base_bdevs_list": [ 00:10:10.634 { 00:10:10.634 "name": "BaseBdev1", 00:10:10.634 "uuid": "7a28cce6-a01f-4951-bac0-f1e60acf918a", 00:10:10.634 "is_configured": true, 00:10:10.634 "data_offset": 2048, 00:10:10.634 "data_size": 63488 00:10:10.634 }, 00:10:10.634 { 00:10:10.634 "name": null, 00:10:10.634 "uuid": "6ff47516-0270-4bd2-845d-fd41575caf37", 00:10:10.634 "is_configured": false, 00:10:10.634 "data_offset": 0, 00:10:10.634 "data_size": 63488 00:10:10.634 }, 00:10:10.634 { 00:10:10.634 "name": "BaseBdev3", 00:10:10.634 "uuid": "fe55b125-51ec-4d2e-a7b8-981e37517c5c", 00:10:10.634 "is_configured": true, 00:10:10.634 "data_offset": 2048, 00:10:10.634 "data_size": 63488 00:10:10.634 } 00:10:10.634 ] 00:10:10.634 }' 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:10.634 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.892 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.892 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:10.892 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.892 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.892 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.892 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:10.892 15:58:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:10.892 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.892 15:58:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.892 [2024-11-20 15:58:08.972545] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:10.892 "name": "Existed_Raid", 00:10:10.892 "uuid": "b0a75242-ace3-4c3c-af0f-79a53721a07a", 00:10:10.892 "strip_size_kb": 64, 00:10:10.892 "state": "configuring", 00:10:10.892 "raid_level": "concat", 00:10:10.892 "superblock": true, 00:10:10.892 "num_base_bdevs": 3, 00:10:10.892 "num_base_bdevs_discovered": 1, 00:10:10.892 "num_base_bdevs_operational": 3, 00:10:10.892 "base_bdevs_list": [ 00:10:10.892 { 00:10:10.892 "name": null, 00:10:10.892 "uuid": "7a28cce6-a01f-4951-bac0-f1e60acf918a", 00:10:10.892 "is_configured": false, 00:10:10.892 "data_offset": 0, 00:10:10.892 "data_size": 63488 00:10:10.892 }, 00:10:10.892 { 00:10:10.892 "name": null, 00:10:10.892 "uuid": "6ff47516-0270-4bd2-845d-fd41575caf37", 00:10:10.892 "is_configured": false, 00:10:10.892 "data_offset": 0, 00:10:10.892 "data_size": 63488 00:10:10.892 }, 00:10:10.892 { 00:10:10.892 "name": "BaseBdev3", 00:10:10.892 "uuid": "fe55b125-51ec-4d2e-a7b8-981e37517c5c", 00:10:10.892 "is_configured": true, 00:10:10.892 "data_offset": 2048, 00:10:10.892 "data_size": 63488 00:10:10.892 } 00:10:10.892 ] 00:10:10.892 }' 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:10.892 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.162 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.162 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.162 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.162 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:11.162 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.162 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:11.162 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:11.162 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.162 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.162 [2024-11-20 15:58:09.354755] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:11.162 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.163 "name": "Existed_Raid", 00:10:11.163 "uuid": "b0a75242-ace3-4c3c-af0f-79a53721a07a", 00:10:11.163 "strip_size_kb": 64, 00:10:11.163 "state": "configuring", 00:10:11.163 "raid_level": "concat", 00:10:11.163 "superblock": true, 00:10:11.163 "num_base_bdevs": 3, 00:10:11.163 "num_base_bdevs_discovered": 2, 00:10:11.163 "num_base_bdevs_operational": 3, 00:10:11.163 "base_bdevs_list": [ 00:10:11.163 { 00:10:11.163 "name": null, 00:10:11.163 "uuid": "7a28cce6-a01f-4951-bac0-f1e60acf918a", 00:10:11.163 "is_configured": false, 00:10:11.163 "data_offset": 0, 00:10:11.163 "data_size": 63488 00:10:11.163 }, 00:10:11.163 { 00:10:11.163 "name": "BaseBdev2", 00:10:11.163 "uuid": "6ff47516-0270-4bd2-845d-fd41575caf37", 00:10:11.163 "is_configured": true, 00:10:11.163 "data_offset": 2048, 00:10:11.163 "data_size": 63488 00:10:11.163 }, 00:10:11.163 { 00:10:11.163 "name": "BaseBdev3", 00:10:11.163 "uuid": "fe55b125-51ec-4d2e-a7b8-981e37517c5c", 00:10:11.163 "is_configured": true, 00:10:11.163 "data_offset": 2048, 00:10:11.163 "data_size": 63488 00:10:11.163 } 00:10:11.163 ] 00:10:11.163 }' 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.163 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 7a28cce6-a01f-4951-bac0-f1e60acf918a 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.732 [2024-11-20 15:58:09.776987] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:11.732 [2024-11-20 15:58:09.777131] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:10:11.732 [2024-11-20 15:58:09.777143] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:10:11.732 NewBaseBdev 00:10:11.732 [2024-11-20 15:58:09.777333] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:10:11.732 [2024-11-20 15:58:09.777431] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:10:11.732 [2024-11-20 15:58:09.777442] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:10:11.732 [2024-11-20 15:58:09.777534] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.732 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.732 [ 00:10:11.732 { 00:10:11.732 "name": "NewBaseBdev", 00:10:11.732 "aliases": [ 00:10:11.732 "7a28cce6-a01f-4951-bac0-f1e60acf918a" 00:10:11.732 ], 00:10:11.732 "product_name": "Malloc disk", 00:10:11.732 "block_size": 512, 00:10:11.733 "num_blocks": 65536, 00:10:11.733 "uuid": "7a28cce6-a01f-4951-bac0-f1e60acf918a", 00:10:11.733 "assigned_rate_limits": { 00:10:11.733 "rw_ios_per_sec": 0, 00:10:11.733 "rw_mbytes_per_sec": 0, 00:10:11.733 "r_mbytes_per_sec": 0, 00:10:11.733 "w_mbytes_per_sec": 0 00:10:11.733 }, 00:10:11.733 "claimed": true, 00:10:11.733 "claim_type": "exclusive_write", 00:10:11.733 "zoned": false, 00:10:11.733 "supported_io_types": { 00:10:11.733 "read": true, 00:10:11.733 "write": true, 00:10:11.733 "unmap": true, 00:10:11.733 "flush": true, 00:10:11.733 "reset": true, 00:10:11.733 "nvme_admin": false, 00:10:11.733 "nvme_io": false, 00:10:11.733 "nvme_io_md": false, 00:10:11.733 "write_zeroes": true, 00:10:11.733 "zcopy": true, 00:10:11.733 "get_zone_info": false, 00:10:11.733 "zone_management": false, 00:10:11.733 "zone_append": false, 00:10:11.733 "compare": false, 00:10:11.733 "compare_and_write": false, 00:10:11.733 "abort": true, 00:10:11.733 "seek_hole": false, 00:10:11.733 "seek_data": false, 00:10:11.733 "copy": true, 00:10:11.733 "nvme_iov_md": false 00:10:11.733 }, 00:10:11.733 "memory_domains": [ 00:10:11.733 { 00:10:11.733 "dma_device_id": "system", 00:10:11.733 "dma_device_type": 1 00:10:11.733 }, 00:10:11.733 { 00:10:11.733 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.733 "dma_device_type": 2 00:10:11.733 } 00:10:11.733 ], 00:10:11.733 "driver_specific": {} 00:10:11.733 } 00:10:11.733 ] 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.733 "name": "Existed_Raid", 00:10:11.733 "uuid": "b0a75242-ace3-4c3c-af0f-79a53721a07a", 00:10:11.733 "strip_size_kb": 64, 00:10:11.733 "state": "online", 00:10:11.733 "raid_level": "concat", 00:10:11.733 "superblock": true, 00:10:11.733 "num_base_bdevs": 3, 00:10:11.733 "num_base_bdevs_discovered": 3, 00:10:11.733 "num_base_bdevs_operational": 3, 00:10:11.733 "base_bdevs_list": [ 00:10:11.733 { 00:10:11.733 "name": "NewBaseBdev", 00:10:11.733 "uuid": "7a28cce6-a01f-4951-bac0-f1e60acf918a", 00:10:11.733 "is_configured": true, 00:10:11.733 "data_offset": 2048, 00:10:11.733 "data_size": 63488 00:10:11.733 }, 00:10:11.733 { 00:10:11.733 "name": "BaseBdev2", 00:10:11.733 "uuid": "6ff47516-0270-4bd2-845d-fd41575caf37", 00:10:11.733 "is_configured": true, 00:10:11.733 "data_offset": 2048, 00:10:11.733 "data_size": 63488 00:10:11.733 }, 00:10:11.733 { 00:10:11.733 "name": "BaseBdev3", 00:10:11.733 "uuid": "fe55b125-51ec-4d2e-a7b8-981e37517c5c", 00:10:11.733 "is_configured": true, 00:10:11.733 "data_offset": 2048, 00:10:11.733 "data_size": 63488 00:10:11.733 } 00:10:11.733 ] 00:10:11.733 }' 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.733 15:58:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.991 [2024-11-20 15:58:10.129324] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:11.991 "name": "Existed_Raid", 00:10:11.991 "aliases": [ 00:10:11.991 "b0a75242-ace3-4c3c-af0f-79a53721a07a" 00:10:11.991 ], 00:10:11.991 "product_name": "Raid Volume", 00:10:11.991 "block_size": 512, 00:10:11.991 "num_blocks": 190464, 00:10:11.991 "uuid": "b0a75242-ace3-4c3c-af0f-79a53721a07a", 00:10:11.991 "assigned_rate_limits": { 00:10:11.991 "rw_ios_per_sec": 0, 00:10:11.991 "rw_mbytes_per_sec": 0, 00:10:11.991 "r_mbytes_per_sec": 0, 00:10:11.991 "w_mbytes_per_sec": 0 00:10:11.991 }, 00:10:11.991 "claimed": false, 00:10:11.991 "zoned": false, 00:10:11.991 "supported_io_types": { 00:10:11.991 "read": true, 00:10:11.991 "write": true, 00:10:11.991 "unmap": true, 00:10:11.991 "flush": true, 00:10:11.991 "reset": true, 00:10:11.991 "nvme_admin": false, 00:10:11.991 "nvme_io": false, 00:10:11.991 "nvme_io_md": false, 00:10:11.991 "write_zeroes": true, 00:10:11.991 "zcopy": false, 00:10:11.991 "get_zone_info": false, 00:10:11.991 "zone_management": false, 00:10:11.991 "zone_append": false, 00:10:11.991 "compare": false, 00:10:11.991 "compare_and_write": false, 00:10:11.991 "abort": false, 00:10:11.991 "seek_hole": false, 00:10:11.991 "seek_data": false, 00:10:11.991 "copy": false, 00:10:11.991 "nvme_iov_md": false 00:10:11.991 }, 00:10:11.991 "memory_domains": [ 00:10:11.991 { 00:10:11.991 "dma_device_id": "system", 00:10:11.991 "dma_device_type": 1 00:10:11.991 }, 00:10:11.991 { 00:10:11.991 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.991 "dma_device_type": 2 00:10:11.991 }, 00:10:11.991 { 00:10:11.991 "dma_device_id": "system", 00:10:11.991 "dma_device_type": 1 00:10:11.991 }, 00:10:11.991 { 00:10:11.991 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.991 "dma_device_type": 2 00:10:11.991 }, 00:10:11.991 { 00:10:11.991 "dma_device_id": "system", 00:10:11.991 "dma_device_type": 1 00:10:11.991 }, 00:10:11.991 { 00:10:11.991 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.991 "dma_device_type": 2 00:10:11.991 } 00:10:11.991 ], 00:10:11.991 "driver_specific": { 00:10:11.991 "raid": { 00:10:11.991 "uuid": "b0a75242-ace3-4c3c-af0f-79a53721a07a", 00:10:11.991 "strip_size_kb": 64, 00:10:11.991 "state": "online", 00:10:11.991 "raid_level": "concat", 00:10:11.991 "superblock": true, 00:10:11.991 "num_base_bdevs": 3, 00:10:11.991 "num_base_bdevs_discovered": 3, 00:10:11.991 "num_base_bdevs_operational": 3, 00:10:11.991 "base_bdevs_list": [ 00:10:11.991 { 00:10:11.991 "name": "NewBaseBdev", 00:10:11.991 "uuid": "7a28cce6-a01f-4951-bac0-f1e60acf918a", 00:10:11.991 "is_configured": true, 00:10:11.991 "data_offset": 2048, 00:10:11.991 "data_size": 63488 00:10:11.991 }, 00:10:11.991 { 00:10:11.991 "name": "BaseBdev2", 00:10:11.991 "uuid": "6ff47516-0270-4bd2-845d-fd41575caf37", 00:10:11.991 "is_configured": true, 00:10:11.991 "data_offset": 2048, 00:10:11.991 "data_size": 63488 00:10:11.991 }, 00:10:11.991 { 00:10:11.991 "name": "BaseBdev3", 00:10:11.991 "uuid": "fe55b125-51ec-4d2e-a7b8-981e37517c5c", 00:10:11.991 "is_configured": true, 00:10:11.991 "data_offset": 2048, 00:10:11.991 "data_size": 63488 00:10:11.991 } 00:10:11.991 ] 00:10:11.991 } 00:10:11.991 } 00:10:11.991 }' 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:11.991 BaseBdev2 00:10:11.991 BaseBdev3' 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:11.991 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.249 [2024-11-20 15:58:10.321098] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:12.249 [2024-11-20 15:58:10.321193] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:12.249 [2024-11-20 15:58:10.321253] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:12.249 [2024-11-20 15:58:10.321301] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:12.249 [2024-11-20 15:58:10.321310] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 64606 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 64606 ']' 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 64606 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:12.249 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 64606 00:10:12.249 killing process with pid 64606 00:10:12.250 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:12.250 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:12.250 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 64606' 00:10:12.250 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 64606 00:10:12.250 [2024-11-20 15:58:10.353840] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:12.250 15:58:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 64606 00:10:12.528 [2024-11-20 15:58:10.501007] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:13.107 15:58:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:13.107 00:10:13.107 real 0m7.336s 00:10:13.107 user 0m11.855s 00:10:13.107 sys 0m1.155s 00:10:13.107 15:58:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:13.107 15:58:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.107 ************************************ 00:10:13.107 END TEST raid_state_function_test_sb 00:10:13.107 ************************************ 00:10:13.107 15:58:11 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 3 00:10:13.107 15:58:11 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:10:13.107 15:58:11 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:13.107 15:58:11 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:13.107 ************************************ 00:10:13.107 START TEST raid_superblock_test 00:10:13.107 ************************************ 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 3 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:13.107 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=65193 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 65193 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 65193 ']' 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:13.107 15:58:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.107 [2024-11-20 15:58:11.174628] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:10:13.107 [2024-11-20 15:58:11.174920] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid65193 ] 00:10:13.107 [2024-11-20 15:58:11.331684] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:13.364 [2024-11-20 15:58:11.415221] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:13.364 [2024-11-20 15:58:11.524695] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:13.364 [2024-11-20 15:58:11.524842] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.929 malloc1 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.929 [2024-11-20 15:58:12.058751] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:13.929 [2024-11-20 15:58:12.058803] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:13.929 [2024-11-20 15:58:12.058820] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:13.929 [2024-11-20 15:58:12.058829] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:13.929 [2024-11-20 15:58:12.060559] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:13.929 [2024-11-20 15:58:12.060589] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:13.929 pt1 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.929 malloc2 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.929 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.929 [2024-11-20 15:58:12.098094] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:13.929 [2024-11-20 15:58:12.098227] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:13.929 [2024-11-20 15:58:12.098250] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:13.929 [2024-11-20 15:58:12.098257] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:13.930 [2024-11-20 15:58:12.099982] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:13.930 [2024-11-20 15:58:12.100008] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:13.930 pt2 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.930 malloc3 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.930 [2024-11-20 15:58:12.144694] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:13.930 [2024-11-20 15:58:12.144740] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:13.930 [2024-11-20 15:58:12.144759] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:13.930 [2024-11-20 15:58:12.144766] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:13.930 [2024-11-20 15:58:12.146518] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:13.930 [2024-11-20 15:58:12.146633] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:13.930 pt3 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.930 [2024-11-20 15:58:12.152745] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:13.930 [2024-11-20 15:58:12.154363] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:13.930 [2024-11-20 15:58:12.154418] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:13.930 [2024-11-20 15:58:12.154569] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:10:13.930 [2024-11-20 15:58:12.154581] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:10:13.930 [2024-11-20 15:58:12.154823] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:10:13.930 [2024-11-20 15:58:12.154953] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:10:13.930 [2024-11-20 15:58:12.154960] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:10:13.930 [2024-11-20 15:58:12.155088] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.930 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.187 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:14.187 "name": "raid_bdev1", 00:10:14.187 "uuid": "58217d25-dbc1-4509-88af-0043911b16c5", 00:10:14.187 "strip_size_kb": 64, 00:10:14.187 "state": "online", 00:10:14.187 "raid_level": "concat", 00:10:14.187 "superblock": true, 00:10:14.187 "num_base_bdevs": 3, 00:10:14.187 "num_base_bdevs_discovered": 3, 00:10:14.187 "num_base_bdevs_operational": 3, 00:10:14.187 "base_bdevs_list": [ 00:10:14.187 { 00:10:14.187 "name": "pt1", 00:10:14.187 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:14.187 "is_configured": true, 00:10:14.187 "data_offset": 2048, 00:10:14.187 "data_size": 63488 00:10:14.187 }, 00:10:14.187 { 00:10:14.187 "name": "pt2", 00:10:14.187 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:14.187 "is_configured": true, 00:10:14.187 "data_offset": 2048, 00:10:14.187 "data_size": 63488 00:10:14.187 }, 00:10:14.187 { 00:10:14.187 "name": "pt3", 00:10:14.187 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:14.187 "is_configured": true, 00:10:14.187 "data_offset": 2048, 00:10:14.187 "data_size": 63488 00:10:14.187 } 00:10:14.187 ] 00:10:14.187 }' 00:10:14.187 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:14.187 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.445 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:14.445 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:14.445 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:14.445 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:14.445 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:14.445 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:14.445 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:14.445 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:14.445 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.445 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.445 [2024-11-20 15:58:12.473039] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:14.445 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.445 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:14.445 "name": "raid_bdev1", 00:10:14.445 "aliases": [ 00:10:14.445 "58217d25-dbc1-4509-88af-0043911b16c5" 00:10:14.445 ], 00:10:14.445 "product_name": "Raid Volume", 00:10:14.445 "block_size": 512, 00:10:14.445 "num_blocks": 190464, 00:10:14.445 "uuid": "58217d25-dbc1-4509-88af-0043911b16c5", 00:10:14.445 "assigned_rate_limits": { 00:10:14.445 "rw_ios_per_sec": 0, 00:10:14.445 "rw_mbytes_per_sec": 0, 00:10:14.445 "r_mbytes_per_sec": 0, 00:10:14.445 "w_mbytes_per_sec": 0 00:10:14.445 }, 00:10:14.445 "claimed": false, 00:10:14.445 "zoned": false, 00:10:14.445 "supported_io_types": { 00:10:14.445 "read": true, 00:10:14.445 "write": true, 00:10:14.445 "unmap": true, 00:10:14.445 "flush": true, 00:10:14.445 "reset": true, 00:10:14.445 "nvme_admin": false, 00:10:14.445 "nvme_io": false, 00:10:14.445 "nvme_io_md": false, 00:10:14.445 "write_zeroes": true, 00:10:14.445 "zcopy": false, 00:10:14.445 "get_zone_info": false, 00:10:14.445 "zone_management": false, 00:10:14.445 "zone_append": false, 00:10:14.445 "compare": false, 00:10:14.445 "compare_and_write": false, 00:10:14.445 "abort": false, 00:10:14.445 "seek_hole": false, 00:10:14.445 "seek_data": false, 00:10:14.445 "copy": false, 00:10:14.445 "nvme_iov_md": false 00:10:14.445 }, 00:10:14.445 "memory_domains": [ 00:10:14.445 { 00:10:14.445 "dma_device_id": "system", 00:10:14.445 "dma_device_type": 1 00:10:14.445 }, 00:10:14.445 { 00:10:14.445 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:14.445 "dma_device_type": 2 00:10:14.445 }, 00:10:14.445 { 00:10:14.445 "dma_device_id": "system", 00:10:14.445 "dma_device_type": 1 00:10:14.445 }, 00:10:14.445 { 00:10:14.445 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:14.445 "dma_device_type": 2 00:10:14.445 }, 00:10:14.445 { 00:10:14.445 "dma_device_id": "system", 00:10:14.445 "dma_device_type": 1 00:10:14.445 }, 00:10:14.445 { 00:10:14.445 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:14.445 "dma_device_type": 2 00:10:14.445 } 00:10:14.445 ], 00:10:14.445 "driver_specific": { 00:10:14.445 "raid": { 00:10:14.445 "uuid": "58217d25-dbc1-4509-88af-0043911b16c5", 00:10:14.445 "strip_size_kb": 64, 00:10:14.445 "state": "online", 00:10:14.445 "raid_level": "concat", 00:10:14.445 "superblock": true, 00:10:14.446 "num_base_bdevs": 3, 00:10:14.446 "num_base_bdevs_discovered": 3, 00:10:14.446 "num_base_bdevs_operational": 3, 00:10:14.446 "base_bdevs_list": [ 00:10:14.446 { 00:10:14.446 "name": "pt1", 00:10:14.446 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:14.446 "is_configured": true, 00:10:14.446 "data_offset": 2048, 00:10:14.446 "data_size": 63488 00:10:14.446 }, 00:10:14.446 { 00:10:14.446 "name": "pt2", 00:10:14.446 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:14.446 "is_configured": true, 00:10:14.446 "data_offset": 2048, 00:10:14.446 "data_size": 63488 00:10:14.446 }, 00:10:14.446 { 00:10:14.446 "name": "pt3", 00:10:14.446 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:14.446 "is_configured": true, 00:10:14.446 "data_offset": 2048, 00:10:14.446 "data_size": 63488 00:10:14.446 } 00:10:14.446 ] 00:10:14.446 } 00:10:14.446 } 00:10:14.446 }' 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:14.446 pt2 00:10:14.446 pt3' 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.446 [2024-11-20 15:58:12.669089] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=58217d25-dbc1-4509-88af-0043911b16c5 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 58217d25-dbc1-4509-88af-0043911b16c5 ']' 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.446 [2024-11-20 15:58:12.688832] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:14.446 [2024-11-20 15:58:12.688853] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:14.446 [2024-11-20 15:58:12.688914] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:14.446 [2024-11-20 15:58:12.688969] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:14.446 [2024-11-20 15:58:12.688977] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:10:14.446 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.704 [2024-11-20 15:58:12.792895] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:14.704 [2024-11-20 15:58:12.794512] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:14.704 [2024-11-20 15:58:12.794555] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:14.704 [2024-11-20 15:58:12.794593] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:14.704 [2024-11-20 15:58:12.794634] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:14.704 [2024-11-20 15:58:12.794650] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:14.704 [2024-11-20 15:58:12.794664] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:14.704 [2024-11-20 15:58:12.794681] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:10:14.704 request: 00:10:14.704 { 00:10:14.704 "name": "raid_bdev1", 00:10:14.704 "raid_level": "concat", 00:10:14.704 "base_bdevs": [ 00:10:14.704 "malloc1", 00:10:14.704 "malloc2", 00:10:14.704 "malloc3" 00:10:14.704 ], 00:10:14.704 "strip_size_kb": 64, 00:10:14.704 "superblock": false, 00:10:14.704 "method": "bdev_raid_create", 00:10:14.704 "req_id": 1 00:10:14.704 } 00:10:14.704 Got JSON-RPC error response 00:10:14.704 response: 00:10:14.704 { 00:10:14.704 "code": -17, 00:10:14.704 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:14.704 } 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.704 [2024-11-20 15:58:12.836863] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:14.704 [2024-11-20 15:58:12.836904] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:14.704 [2024-11-20 15:58:12.836919] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:14.704 [2024-11-20 15:58:12.836926] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:14.704 [2024-11-20 15:58:12.838740] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:14.704 [2024-11-20 15:58:12.838766] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:14.704 [2024-11-20 15:58:12.838827] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:14.704 [2024-11-20 15:58:12.838864] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:14.704 pt1 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.704 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:14.704 "name": "raid_bdev1", 00:10:14.704 "uuid": "58217d25-dbc1-4509-88af-0043911b16c5", 00:10:14.704 "strip_size_kb": 64, 00:10:14.704 "state": "configuring", 00:10:14.704 "raid_level": "concat", 00:10:14.704 "superblock": true, 00:10:14.704 "num_base_bdevs": 3, 00:10:14.704 "num_base_bdevs_discovered": 1, 00:10:14.704 "num_base_bdevs_operational": 3, 00:10:14.704 "base_bdevs_list": [ 00:10:14.704 { 00:10:14.704 "name": "pt1", 00:10:14.705 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:14.705 "is_configured": true, 00:10:14.705 "data_offset": 2048, 00:10:14.705 "data_size": 63488 00:10:14.705 }, 00:10:14.705 { 00:10:14.705 "name": null, 00:10:14.705 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:14.705 "is_configured": false, 00:10:14.705 "data_offset": 2048, 00:10:14.705 "data_size": 63488 00:10:14.705 }, 00:10:14.705 { 00:10:14.705 "name": null, 00:10:14.705 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:14.705 "is_configured": false, 00:10:14.705 "data_offset": 2048, 00:10:14.705 "data_size": 63488 00:10:14.705 } 00:10:14.705 ] 00:10:14.705 }' 00:10:14.705 15:58:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:14.705 15:58:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.963 [2024-11-20 15:58:13.160956] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:14.963 [2024-11-20 15:58:13.161015] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:14.963 [2024-11-20 15:58:13.161035] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:10:14.963 [2024-11-20 15:58:13.161042] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:14.963 [2024-11-20 15:58:13.161379] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:14.963 [2024-11-20 15:58:13.161391] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:14.963 [2024-11-20 15:58:13.161454] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:14.963 [2024-11-20 15:58:13.161473] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:14.963 pt2 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.963 [2024-11-20 15:58:13.168957] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:14.963 "name": "raid_bdev1", 00:10:14.963 "uuid": "58217d25-dbc1-4509-88af-0043911b16c5", 00:10:14.963 "strip_size_kb": 64, 00:10:14.963 "state": "configuring", 00:10:14.963 "raid_level": "concat", 00:10:14.963 "superblock": true, 00:10:14.963 "num_base_bdevs": 3, 00:10:14.963 "num_base_bdevs_discovered": 1, 00:10:14.963 "num_base_bdevs_operational": 3, 00:10:14.963 "base_bdevs_list": [ 00:10:14.963 { 00:10:14.963 "name": "pt1", 00:10:14.963 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:14.963 "is_configured": true, 00:10:14.963 "data_offset": 2048, 00:10:14.963 "data_size": 63488 00:10:14.963 }, 00:10:14.963 { 00:10:14.963 "name": null, 00:10:14.963 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:14.963 "is_configured": false, 00:10:14.963 "data_offset": 0, 00:10:14.963 "data_size": 63488 00:10:14.963 }, 00:10:14.963 { 00:10:14.963 "name": null, 00:10:14.963 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:14.963 "is_configured": false, 00:10:14.963 "data_offset": 2048, 00:10:14.963 "data_size": 63488 00:10:14.963 } 00:10:14.963 ] 00:10:14.963 }' 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:14.963 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.530 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:15.530 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:15.530 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:15.530 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.530 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.530 [2024-11-20 15:58:13.485011] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:15.530 [2024-11-20 15:58:13.485069] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:15.530 [2024-11-20 15:58:13.485084] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:10:15.530 [2024-11-20 15:58:13.485093] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:15.530 [2024-11-20 15:58:13.485452] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:15.530 [2024-11-20 15:58:13.485477] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:15.530 [2024-11-20 15:58:13.485541] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:15.530 [2024-11-20 15:58:13.485560] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:15.530 pt2 00:10:15.530 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.530 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:15.530 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:15.530 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:15.530 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.530 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.530 [2024-11-20 15:58:13.493008] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:15.530 [2024-11-20 15:58:13.493053] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:15.530 [2024-11-20 15:58:13.493066] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:15.530 [2024-11-20 15:58:13.493074] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:15.530 [2024-11-20 15:58:13.493391] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:15.530 [2024-11-20 15:58:13.493411] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:15.530 [2024-11-20 15:58:13.493464] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:15.530 [2024-11-20 15:58:13.493482] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:15.530 [2024-11-20 15:58:13.493575] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:10:15.530 [2024-11-20 15:58:13.493584] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:10:15.531 [2024-11-20 15:58:13.493796] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:10:15.531 [2024-11-20 15:58:13.493905] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:10:15.531 [2024-11-20 15:58:13.493911] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:10:15.531 [2024-11-20 15:58:13.494016] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:15.531 pt3 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:15.531 "name": "raid_bdev1", 00:10:15.531 "uuid": "58217d25-dbc1-4509-88af-0043911b16c5", 00:10:15.531 "strip_size_kb": 64, 00:10:15.531 "state": "online", 00:10:15.531 "raid_level": "concat", 00:10:15.531 "superblock": true, 00:10:15.531 "num_base_bdevs": 3, 00:10:15.531 "num_base_bdevs_discovered": 3, 00:10:15.531 "num_base_bdevs_operational": 3, 00:10:15.531 "base_bdevs_list": [ 00:10:15.531 { 00:10:15.531 "name": "pt1", 00:10:15.531 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:15.531 "is_configured": true, 00:10:15.531 "data_offset": 2048, 00:10:15.531 "data_size": 63488 00:10:15.531 }, 00:10:15.531 { 00:10:15.531 "name": "pt2", 00:10:15.531 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:15.531 "is_configured": true, 00:10:15.531 "data_offset": 2048, 00:10:15.531 "data_size": 63488 00:10:15.531 }, 00:10:15.531 { 00:10:15.531 "name": "pt3", 00:10:15.531 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:15.531 "is_configured": true, 00:10:15.531 "data_offset": 2048, 00:10:15.531 "data_size": 63488 00:10:15.531 } 00:10:15.531 ] 00:10:15.531 }' 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:15.531 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:15.789 [2024-11-20 15:58:13.805353] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:15.789 "name": "raid_bdev1", 00:10:15.789 "aliases": [ 00:10:15.789 "58217d25-dbc1-4509-88af-0043911b16c5" 00:10:15.789 ], 00:10:15.789 "product_name": "Raid Volume", 00:10:15.789 "block_size": 512, 00:10:15.789 "num_blocks": 190464, 00:10:15.789 "uuid": "58217d25-dbc1-4509-88af-0043911b16c5", 00:10:15.789 "assigned_rate_limits": { 00:10:15.789 "rw_ios_per_sec": 0, 00:10:15.789 "rw_mbytes_per_sec": 0, 00:10:15.789 "r_mbytes_per_sec": 0, 00:10:15.789 "w_mbytes_per_sec": 0 00:10:15.789 }, 00:10:15.789 "claimed": false, 00:10:15.789 "zoned": false, 00:10:15.789 "supported_io_types": { 00:10:15.789 "read": true, 00:10:15.789 "write": true, 00:10:15.789 "unmap": true, 00:10:15.789 "flush": true, 00:10:15.789 "reset": true, 00:10:15.789 "nvme_admin": false, 00:10:15.789 "nvme_io": false, 00:10:15.789 "nvme_io_md": false, 00:10:15.789 "write_zeroes": true, 00:10:15.789 "zcopy": false, 00:10:15.789 "get_zone_info": false, 00:10:15.789 "zone_management": false, 00:10:15.789 "zone_append": false, 00:10:15.789 "compare": false, 00:10:15.789 "compare_and_write": false, 00:10:15.789 "abort": false, 00:10:15.789 "seek_hole": false, 00:10:15.789 "seek_data": false, 00:10:15.789 "copy": false, 00:10:15.789 "nvme_iov_md": false 00:10:15.789 }, 00:10:15.789 "memory_domains": [ 00:10:15.789 { 00:10:15.789 "dma_device_id": "system", 00:10:15.789 "dma_device_type": 1 00:10:15.789 }, 00:10:15.789 { 00:10:15.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:15.789 "dma_device_type": 2 00:10:15.789 }, 00:10:15.789 { 00:10:15.789 "dma_device_id": "system", 00:10:15.789 "dma_device_type": 1 00:10:15.789 }, 00:10:15.789 { 00:10:15.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:15.789 "dma_device_type": 2 00:10:15.789 }, 00:10:15.789 { 00:10:15.789 "dma_device_id": "system", 00:10:15.789 "dma_device_type": 1 00:10:15.789 }, 00:10:15.789 { 00:10:15.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:15.789 "dma_device_type": 2 00:10:15.789 } 00:10:15.789 ], 00:10:15.789 "driver_specific": { 00:10:15.789 "raid": { 00:10:15.789 "uuid": "58217d25-dbc1-4509-88af-0043911b16c5", 00:10:15.789 "strip_size_kb": 64, 00:10:15.789 "state": "online", 00:10:15.789 "raid_level": "concat", 00:10:15.789 "superblock": true, 00:10:15.789 "num_base_bdevs": 3, 00:10:15.789 "num_base_bdevs_discovered": 3, 00:10:15.789 "num_base_bdevs_operational": 3, 00:10:15.789 "base_bdevs_list": [ 00:10:15.789 { 00:10:15.789 "name": "pt1", 00:10:15.789 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:15.789 "is_configured": true, 00:10:15.789 "data_offset": 2048, 00:10:15.789 "data_size": 63488 00:10:15.789 }, 00:10:15.789 { 00:10:15.789 "name": "pt2", 00:10:15.789 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:15.789 "is_configured": true, 00:10:15.789 "data_offset": 2048, 00:10:15.789 "data_size": 63488 00:10:15.789 }, 00:10:15.789 { 00:10:15.789 "name": "pt3", 00:10:15.789 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:15.789 "is_configured": true, 00:10:15.789 "data_offset": 2048, 00:10:15.789 "data_size": 63488 00:10:15.789 } 00:10:15.789 ] 00:10:15.789 } 00:10:15.789 } 00:10:15.789 }' 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:15.789 pt2 00:10:15.789 pt3' 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:15.789 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.790 15:58:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.790 15:58:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:15.790 [2024-11-20 15:58:13.993371] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:15.790 15:58:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.790 15:58:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 58217d25-dbc1-4509-88af-0043911b16c5 '!=' 58217d25-dbc1-4509-88af-0043911b16c5 ']' 00:10:15.790 15:58:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:10:15.790 15:58:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:15.790 15:58:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:15.790 15:58:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 65193 00:10:15.790 15:58:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 65193 ']' 00:10:15.790 15:58:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 65193 00:10:15.790 15:58:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:10:15.790 15:58:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:15.790 15:58:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 65193 00:10:16.048 killing process with pid 65193 00:10:16.048 15:58:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:16.048 15:58:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:16.048 15:58:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 65193' 00:10:16.048 15:58:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 65193 00:10:16.048 [2024-11-20 15:58:14.043423] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:16.048 [2024-11-20 15:58:14.043495] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:16.048 15:58:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 65193 00:10:16.048 [2024-11-20 15:58:14.043545] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:16.048 [2024-11-20 15:58:14.043555] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:10:16.048 [2024-11-20 15:58:14.193729] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:16.648 15:58:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:16.648 00:10:16.648 real 0m3.660s 00:10:16.648 user 0m5.337s 00:10:16.648 sys 0m0.579s 00:10:16.648 15:58:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:16.648 15:58:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.648 ************************************ 00:10:16.648 END TEST raid_superblock_test 00:10:16.648 ************************************ 00:10:16.648 15:58:14 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 3 read 00:10:16.648 15:58:14 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:16.648 15:58:14 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:16.648 15:58:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:16.649 ************************************ 00:10:16.649 START TEST raid_read_error_test 00:10:16.649 ************************************ 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 3 read 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.RnzaeUMxUa 00:10:16.649 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=65435 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 65435 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 65435 ']' 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:16.649 15:58:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.649 [2024-11-20 15:58:14.896032] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:10:16.649 [2024-11-20 15:58:14.897329] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid65435 ] 00:10:16.907 [2024-11-20 15:58:15.063716] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:16.907 [2024-11-20 15:58:15.148257] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:17.164 [2024-11-20 15:58:15.258140] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:17.164 [2024-11-20 15:58:15.258186] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.730 BaseBdev1_malloc 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.730 true 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.730 [2024-11-20 15:58:15.728293] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:17.730 [2024-11-20 15:58:15.728340] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:17.730 [2024-11-20 15:58:15.728356] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:10:17.730 [2024-11-20 15:58:15.728365] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:17.730 [2024-11-20 15:58:15.730131] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:17.730 [2024-11-20 15:58:15.730163] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:17.730 BaseBdev1 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.730 BaseBdev2_malloc 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.730 true 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.730 [2024-11-20 15:58:15.767485] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:17.730 [2024-11-20 15:58:15.767524] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:17.730 [2024-11-20 15:58:15.767536] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:17.730 [2024-11-20 15:58:15.767544] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:17.730 [2024-11-20 15:58:15.769265] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:17.730 [2024-11-20 15:58:15.769294] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:17.730 BaseBdev2 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.730 BaseBdev3_malloc 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.730 true 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.730 [2024-11-20 15:58:15.821083] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:17.730 [2024-11-20 15:58:15.821128] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:17.730 [2024-11-20 15:58:15.821142] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:17.730 [2024-11-20 15:58:15.821151] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:17.730 [2024-11-20 15:58:15.822916] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:17.730 [2024-11-20 15:58:15.823072] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:17.730 BaseBdev3 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.730 [2024-11-20 15:58:15.829157] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:17.730 [2024-11-20 15:58:15.830692] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:17.730 [2024-11-20 15:58:15.830752] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:17.730 [2024-11-20 15:58:15.830911] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:10:17.730 [2024-11-20 15:58:15.830920] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:10:17.730 [2024-11-20 15:58:15.831129] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:10:17.730 [2024-11-20 15:58:15.831243] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:10:17.730 [2024-11-20 15:58:15.831254] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:10:17.730 [2024-11-20 15:58:15.831364] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.730 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:17.730 "name": "raid_bdev1", 00:10:17.730 "uuid": "4e161e75-932c-4453-b01c-6b1948a825a5", 00:10:17.730 "strip_size_kb": 64, 00:10:17.730 "state": "online", 00:10:17.730 "raid_level": "concat", 00:10:17.730 "superblock": true, 00:10:17.730 "num_base_bdevs": 3, 00:10:17.730 "num_base_bdevs_discovered": 3, 00:10:17.730 "num_base_bdevs_operational": 3, 00:10:17.730 "base_bdevs_list": [ 00:10:17.730 { 00:10:17.730 "name": "BaseBdev1", 00:10:17.730 "uuid": "25b7ab96-38fc-5766-8aab-8d411138d7a7", 00:10:17.730 "is_configured": true, 00:10:17.730 "data_offset": 2048, 00:10:17.730 "data_size": 63488 00:10:17.730 }, 00:10:17.730 { 00:10:17.730 "name": "BaseBdev2", 00:10:17.730 "uuid": "dab43a04-e94b-540a-ae96-d29d0a10a24b", 00:10:17.730 "is_configured": true, 00:10:17.730 "data_offset": 2048, 00:10:17.730 "data_size": 63488 00:10:17.730 }, 00:10:17.730 { 00:10:17.730 "name": "BaseBdev3", 00:10:17.730 "uuid": "e86a4d98-2485-540a-a0e4-c09a56977e5f", 00:10:17.730 "is_configured": true, 00:10:17.730 "data_offset": 2048, 00:10:17.730 "data_size": 63488 00:10:17.730 } 00:10:17.730 ] 00:10:17.730 }' 00:10:17.731 15:58:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:17.731 15:58:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.988 15:58:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:17.988 15:58:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:17.988 [2024-11-20 15:58:16.225996] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:10:18.920 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:18.920 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.920 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.921 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.178 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:19.178 "name": "raid_bdev1", 00:10:19.178 "uuid": "4e161e75-932c-4453-b01c-6b1948a825a5", 00:10:19.178 "strip_size_kb": 64, 00:10:19.178 "state": "online", 00:10:19.178 "raid_level": "concat", 00:10:19.178 "superblock": true, 00:10:19.178 "num_base_bdevs": 3, 00:10:19.178 "num_base_bdevs_discovered": 3, 00:10:19.178 "num_base_bdevs_operational": 3, 00:10:19.178 "base_bdevs_list": [ 00:10:19.178 { 00:10:19.178 "name": "BaseBdev1", 00:10:19.178 "uuid": "25b7ab96-38fc-5766-8aab-8d411138d7a7", 00:10:19.178 "is_configured": true, 00:10:19.178 "data_offset": 2048, 00:10:19.178 "data_size": 63488 00:10:19.178 }, 00:10:19.179 { 00:10:19.179 "name": "BaseBdev2", 00:10:19.179 "uuid": "dab43a04-e94b-540a-ae96-d29d0a10a24b", 00:10:19.179 "is_configured": true, 00:10:19.179 "data_offset": 2048, 00:10:19.179 "data_size": 63488 00:10:19.179 }, 00:10:19.179 { 00:10:19.179 "name": "BaseBdev3", 00:10:19.179 "uuid": "e86a4d98-2485-540a-a0e4-c09a56977e5f", 00:10:19.179 "is_configured": true, 00:10:19.179 "data_offset": 2048, 00:10:19.179 "data_size": 63488 00:10:19.179 } 00:10:19.179 ] 00:10:19.179 }' 00:10:19.179 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:19.179 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.437 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:19.437 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.437 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.437 [2024-11-20 15:58:17.465649] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:19.437 [2024-11-20 15:58:17.465797] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:19.437 [2024-11-20 15:58:17.468310] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:19.437 [2024-11-20 15:58:17.468424] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:19.437 [2024-11-20 15:58:17.468477] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:19.437 [2024-11-20 15:58:17.468632] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:10:19.437 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.437 { 00:10:19.437 "results": [ 00:10:19.437 { 00:10:19.437 "job": "raid_bdev1", 00:10:19.437 "core_mask": "0x1", 00:10:19.437 "workload": "randrw", 00:10:19.437 "percentage": 50, 00:10:19.437 "status": "finished", 00:10:19.437 "queue_depth": 1, 00:10:19.437 "io_size": 131072, 00:10:19.437 "runtime": 1.238215, 00:10:19.437 "iops": 17902.383673271604, 00:10:19.437 "mibps": 2237.7979591589506, 00:10:19.437 "io_failed": 1, 00:10:19.437 "io_timeout": 0, 00:10:19.437 "avg_latency_us": 76.28935735502318, 00:10:19.437 "min_latency_us": 26.38769230769231, 00:10:19.437 "max_latency_us": 1361.1323076923077 00:10:19.437 } 00:10:19.437 ], 00:10:19.437 "core_count": 1 00:10:19.437 } 00:10:19.437 15:58:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 65435 00:10:19.437 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 65435 ']' 00:10:19.437 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 65435 00:10:19.437 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:10:19.437 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:19.437 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 65435 00:10:19.437 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:19.437 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:19.437 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 65435' 00:10:19.437 killing process with pid 65435 00:10:19.437 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 65435 00:10:19.437 [2024-11-20 15:58:17.495447] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:19.437 15:58:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 65435 00:10:19.437 [2024-11-20 15:58:17.607168] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:20.003 15:58:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.RnzaeUMxUa 00:10:20.003 15:58:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:20.003 15:58:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:20.003 15:58:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.81 00:10:20.003 15:58:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:10:20.003 15:58:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:20.003 15:58:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:20.003 15:58:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.81 != \0\.\0\0 ]] 00:10:20.003 00:10:20.003 real 0m3.391s 00:10:20.003 user 0m4.028s 00:10:20.003 sys 0m0.383s 00:10:20.003 15:58:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:20.003 15:58:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.003 ************************************ 00:10:20.003 END TEST raid_read_error_test 00:10:20.003 ************************************ 00:10:20.003 15:58:18 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 3 write 00:10:20.003 15:58:18 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:20.003 15:58:18 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:20.003 15:58:18 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:20.003 ************************************ 00:10:20.003 START TEST raid_write_error_test 00:10:20.003 ************************************ 00:10:20.003 15:58:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 3 write 00:10:20.003 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:20.003 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:10:20.003 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:20.261 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:20.262 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:20.262 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:20.262 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:20.262 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:20.262 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.UbGVe9Q3rK 00:10:20.262 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=65564 00:10:20.262 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 65564 00:10:20.262 15:58:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 65564 ']' 00:10:20.262 15:58:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:20.262 15:58:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:20.262 15:58:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:20.262 15:58:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:20.262 15:58:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.262 15:58:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:20.262 [2024-11-20 15:58:18.322561] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:10:20.262 [2024-11-20 15:58:18.322867] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid65564 ] 00:10:20.262 [2024-11-20 15:58:18.474675] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:20.521 [2024-11-20 15:58:18.574294] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:20.521 [2024-11-20 15:58:18.710254] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:20.521 [2024-11-20 15:58:18.710293] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:21.093 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:21.093 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:21.093 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:21.093 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:21.093 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.093 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.093 BaseBdev1_malloc 00:10:21.093 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.093 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:21.093 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.093 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.093 true 00:10:21.093 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.093 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:21.093 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.093 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.093 [2024-11-20 15:58:19.216730] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:21.093 [2024-11-20 15:58:19.216904] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:21.093 [2024-11-20 15:58:19.216929] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:10:21.093 [2024-11-20 15:58:19.216940] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:21.093 [2024-11-20 15:58:19.219035] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:21.093 [2024-11-20 15:58:19.219071] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:21.093 BaseBdev1 00:10:21.093 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.093 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:21.094 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:21.094 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.094 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.094 BaseBdev2_malloc 00:10:21.094 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.094 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:21.094 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.094 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.094 true 00:10:21.094 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.094 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:21.094 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.094 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.095 [2024-11-20 15:58:19.260438] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:21.095 [2024-11-20 15:58:19.260601] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:21.095 [2024-11-20 15:58:19.260623] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:21.095 [2024-11-20 15:58:19.260633] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:21.095 [2024-11-20 15:58:19.262716] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:21.095 [2024-11-20 15:58:19.262745] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:21.095 BaseBdev2 00:10:21.095 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.095 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:21.095 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:21.095 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.095 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.095 BaseBdev3_malloc 00:10:21.095 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.095 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:21.095 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.095 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.095 true 00:10:21.095 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.095 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:21.095 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.095 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.095 [2024-11-20 15:58:19.314656] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:21.096 [2024-11-20 15:58:19.314715] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:21.096 [2024-11-20 15:58:19.314732] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:21.096 [2024-11-20 15:58:19.314743] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:21.096 [2024-11-20 15:58:19.316876] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:21.096 [2024-11-20 15:58:19.316911] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:21.096 BaseBdev3 00:10:21.096 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.096 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:10:21.096 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.096 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.096 [2024-11-20 15:58:19.322728] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:21.096 [2024-11-20 15:58:19.324540] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:21.096 [2024-11-20 15:58:19.324614] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:21.096 [2024-11-20 15:58:19.324817] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:10:21.096 [2024-11-20 15:58:19.324828] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:10:21.096 [2024-11-20 15:58:19.325077] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:10:21.096 [2024-11-20 15:58:19.325217] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:10:21.096 [2024-11-20 15:58:19.325229] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:10:21.096 [2024-11-20 15:58:19.325360] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:21.096 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.096 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:10:21.100 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:21.100 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:21.100 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:21.100 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:21.100 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:21.100 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:21.100 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:21.101 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:21.101 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:21.101 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.101 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:21.101 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.101 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.364 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.364 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:21.364 "name": "raid_bdev1", 00:10:21.364 "uuid": "2180bf97-a353-4a64-b5fa-0a48e2747a4f", 00:10:21.364 "strip_size_kb": 64, 00:10:21.364 "state": "online", 00:10:21.364 "raid_level": "concat", 00:10:21.364 "superblock": true, 00:10:21.364 "num_base_bdevs": 3, 00:10:21.364 "num_base_bdevs_discovered": 3, 00:10:21.364 "num_base_bdevs_operational": 3, 00:10:21.364 "base_bdevs_list": [ 00:10:21.364 { 00:10:21.364 "name": "BaseBdev1", 00:10:21.364 "uuid": "75c9bc0a-25f1-5459-825c-10687a5f4dd7", 00:10:21.364 "is_configured": true, 00:10:21.364 "data_offset": 2048, 00:10:21.364 "data_size": 63488 00:10:21.364 }, 00:10:21.364 { 00:10:21.364 "name": "BaseBdev2", 00:10:21.364 "uuid": "fd55fe82-5ec1-5a36-a1f7-6d73eb949ff7", 00:10:21.364 "is_configured": true, 00:10:21.364 "data_offset": 2048, 00:10:21.364 "data_size": 63488 00:10:21.364 }, 00:10:21.364 { 00:10:21.364 "name": "BaseBdev3", 00:10:21.364 "uuid": "f10e87d2-02b9-55ad-a4fc-dc277cb05815", 00:10:21.364 "is_configured": true, 00:10:21.364 "data_offset": 2048, 00:10:21.364 "data_size": 63488 00:10:21.364 } 00:10:21.364 ] 00:10:21.364 }' 00:10:21.364 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:21.364 15:58:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.621 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:21.621 15:58:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:21.621 [2024-11-20 15:58:19.731737] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:22.554 "name": "raid_bdev1", 00:10:22.554 "uuid": "2180bf97-a353-4a64-b5fa-0a48e2747a4f", 00:10:22.554 "strip_size_kb": 64, 00:10:22.554 "state": "online", 00:10:22.554 "raid_level": "concat", 00:10:22.554 "superblock": true, 00:10:22.554 "num_base_bdevs": 3, 00:10:22.554 "num_base_bdevs_discovered": 3, 00:10:22.554 "num_base_bdevs_operational": 3, 00:10:22.554 "base_bdevs_list": [ 00:10:22.554 { 00:10:22.554 "name": "BaseBdev1", 00:10:22.554 "uuid": "75c9bc0a-25f1-5459-825c-10687a5f4dd7", 00:10:22.554 "is_configured": true, 00:10:22.554 "data_offset": 2048, 00:10:22.554 "data_size": 63488 00:10:22.554 }, 00:10:22.554 { 00:10:22.554 "name": "BaseBdev2", 00:10:22.554 "uuid": "fd55fe82-5ec1-5a36-a1f7-6d73eb949ff7", 00:10:22.554 "is_configured": true, 00:10:22.554 "data_offset": 2048, 00:10:22.554 "data_size": 63488 00:10:22.554 }, 00:10:22.554 { 00:10:22.554 "name": "BaseBdev3", 00:10:22.554 "uuid": "f10e87d2-02b9-55ad-a4fc-dc277cb05815", 00:10:22.554 "is_configured": true, 00:10:22.554 "data_offset": 2048, 00:10:22.554 "data_size": 63488 00:10:22.554 } 00:10:22.554 ] 00:10:22.554 }' 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:22.554 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.812 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:22.812 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.812 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.812 [2024-11-20 15:58:20.961661] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:22.812 [2024-11-20 15:58:20.961702] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:22.812 [2024-11-20 15:58:20.964954] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:22.812 [2024-11-20 15:58:20.965093] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:22.812 [2024-11-20 15:58:20.965155] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:22.812 [2024-11-20 15:58:20.965497] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:10:22.812 { 00:10:22.812 "results": [ 00:10:22.812 { 00:10:22.812 "job": "raid_bdev1", 00:10:22.812 "core_mask": "0x1", 00:10:22.812 "workload": "randrw", 00:10:22.812 "percentage": 50, 00:10:22.812 "status": "finished", 00:10:22.812 "queue_depth": 1, 00:10:22.812 "io_size": 131072, 00:10:22.812 "runtime": 1.228171, 00:10:22.812 "iops": 14716.191800653167, 00:10:22.812 "mibps": 1839.5239750816459, 00:10:22.812 "io_failed": 1, 00:10:22.812 "io_timeout": 0, 00:10:22.812 "avg_latency_us": 92.5828099159485, 00:10:22.812 "min_latency_us": 33.673846153846156, 00:10:22.812 "max_latency_us": 1751.8276923076924 00:10:22.812 } 00:10:22.812 ], 00:10:22.812 "core_count": 1 00:10:22.812 } 00:10:22.812 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.812 15:58:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 65564 00:10:22.812 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 65564 ']' 00:10:22.812 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 65564 00:10:22.812 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:10:22.812 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:22.812 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 65564 00:10:22.812 killing process with pid 65564 00:10:22.812 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:22.812 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:22.812 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 65564' 00:10:22.812 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 65564 00:10:22.812 [2024-11-20 15:58:20.995258] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:22.812 15:58:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 65564 00:10:23.070 [2024-11-20 15:58:21.136831] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:23.689 15:58:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:23.690 15:58:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.UbGVe9Q3rK 00:10:23.690 15:58:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:23.690 ************************************ 00:10:23.690 END TEST raid_write_error_test 00:10:23.690 ************************************ 00:10:23.690 15:58:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.81 00:10:23.690 15:58:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:10:23.690 15:58:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:23.690 15:58:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:23.690 15:58:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.81 != \0\.\0\0 ]] 00:10:23.690 00:10:23.690 real 0m3.646s 00:10:23.690 user 0m4.344s 00:10:23.690 sys 0m0.383s 00:10:23.690 15:58:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:23.690 15:58:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.690 15:58:21 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:10:23.690 15:58:21 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 3 false 00:10:23.690 15:58:21 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:23.690 15:58:21 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:23.690 15:58:21 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:23.690 ************************************ 00:10:23.690 START TEST raid_state_function_test 00:10:23.690 ************************************ 00:10:23.690 15:58:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 3 false 00:10:23.690 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:10:23.690 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:10:23.690 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:10:23.690 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:23.690 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:23.690 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:23.947 Process raid pid: 65701 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=65701 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 65701' 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 65701 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 65701 ']' 00:10:23.947 15:58:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:23.948 15:58:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:23.948 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:23.948 15:58:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:23.948 15:58:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:23.948 15:58:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.948 15:58:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:23.948 [2024-11-20 15:58:22.003728] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:10:23.948 [2024-11-20 15:58:22.003848] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:23.948 [2024-11-20 15:58:22.155608] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:24.205 [2024-11-20 15:58:22.254588] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:24.205 [2024-11-20 15:58:22.390085] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:24.205 [2024-11-20 15:58:22.390118] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.770 [2024-11-20 15:58:22.851176] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:24.770 [2024-11-20 15:58:22.851225] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:24.770 [2024-11-20 15:58:22.851235] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:24.770 [2024-11-20 15:58:22.851244] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:24.770 [2024-11-20 15:58:22.851250] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:24.770 [2024-11-20 15:58:22.851258] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:24.770 "name": "Existed_Raid", 00:10:24.770 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.770 "strip_size_kb": 0, 00:10:24.770 "state": "configuring", 00:10:24.770 "raid_level": "raid1", 00:10:24.770 "superblock": false, 00:10:24.770 "num_base_bdevs": 3, 00:10:24.770 "num_base_bdevs_discovered": 0, 00:10:24.770 "num_base_bdevs_operational": 3, 00:10:24.770 "base_bdevs_list": [ 00:10:24.770 { 00:10:24.770 "name": "BaseBdev1", 00:10:24.770 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.770 "is_configured": false, 00:10:24.770 "data_offset": 0, 00:10:24.770 "data_size": 0 00:10:24.770 }, 00:10:24.770 { 00:10:24.770 "name": "BaseBdev2", 00:10:24.770 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.770 "is_configured": false, 00:10:24.770 "data_offset": 0, 00:10:24.770 "data_size": 0 00:10:24.770 }, 00:10:24.770 { 00:10:24.770 "name": "BaseBdev3", 00:10:24.770 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.770 "is_configured": false, 00:10:24.770 "data_offset": 0, 00:10:24.770 "data_size": 0 00:10:24.770 } 00:10:24.770 ] 00:10:24.770 }' 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:24.770 15:58:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.028 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:25.028 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.028 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.028 [2024-11-20 15:58:23.159180] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:25.028 [2024-11-20 15:58:23.159207] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:10:25.028 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.028 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:25.028 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.028 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.028 [2024-11-20 15:58:23.167180] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:25.028 [2024-11-20 15:58:23.167214] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:25.028 [2024-11-20 15:58:23.167221] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:25.028 [2024-11-20 15:58:23.167228] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:25.028 [2024-11-20 15:58:23.167232] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:25.028 [2024-11-20 15:58:23.167239] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:25.028 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.028 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:25.028 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.029 [2024-11-20 15:58:23.194938] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:25.029 BaseBdev1 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.029 [ 00:10:25.029 { 00:10:25.029 "name": "BaseBdev1", 00:10:25.029 "aliases": [ 00:10:25.029 "9dfe0b7c-5314-4793-a938-e488633e26a1" 00:10:25.029 ], 00:10:25.029 "product_name": "Malloc disk", 00:10:25.029 "block_size": 512, 00:10:25.029 "num_blocks": 65536, 00:10:25.029 "uuid": "9dfe0b7c-5314-4793-a938-e488633e26a1", 00:10:25.029 "assigned_rate_limits": { 00:10:25.029 "rw_ios_per_sec": 0, 00:10:25.029 "rw_mbytes_per_sec": 0, 00:10:25.029 "r_mbytes_per_sec": 0, 00:10:25.029 "w_mbytes_per_sec": 0 00:10:25.029 }, 00:10:25.029 "claimed": true, 00:10:25.029 "claim_type": "exclusive_write", 00:10:25.029 "zoned": false, 00:10:25.029 "supported_io_types": { 00:10:25.029 "read": true, 00:10:25.029 "write": true, 00:10:25.029 "unmap": true, 00:10:25.029 "flush": true, 00:10:25.029 "reset": true, 00:10:25.029 "nvme_admin": false, 00:10:25.029 "nvme_io": false, 00:10:25.029 "nvme_io_md": false, 00:10:25.029 "write_zeroes": true, 00:10:25.029 "zcopy": true, 00:10:25.029 "get_zone_info": false, 00:10:25.029 "zone_management": false, 00:10:25.029 "zone_append": false, 00:10:25.029 "compare": false, 00:10:25.029 "compare_and_write": false, 00:10:25.029 "abort": true, 00:10:25.029 "seek_hole": false, 00:10:25.029 "seek_data": false, 00:10:25.029 "copy": true, 00:10:25.029 "nvme_iov_md": false 00:10:25.029 }, 00:10:25.029 "memory_domains": [ 00:10:25.029 { 00:10:25.029 "dma_device_id": "system", 00:10:25.029 "dma_device_type": 1 00:10:25.029 }, 00:10:25.029 { 00:10:25.029 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:25.029 "dma_device_type": 2 00:10:25.029 } 00:10:25.029 ], 00:10:25.029 "driver_specific": {} 00:10:25.029 } 00:10:25.029 ] 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:25.029 "name": "Existed_Raid", 00:10:25.029 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.029 "strip_size_kb": 0, 00:10:25.029 "state": "configuring", 00:10:25.029 "raid_level": "raid1", 00:10:25.029 "superblock": false, 00:10:25.029 "num_base_bdevs": 3, 00:10:25.029 "num_base_bdevs_discovered": 1, 00:10:25.029 "num_base_bdevs_operational": 3, 00:10:25.029 "base_bdevs_list": [ 00:10:25.029 { 00:10:25.029 "name": "BaseBdev1", 00:10:25.029 "uuid": "9dfe0b7c-5314-4793-a938-e488633e26a1", 00:10:25.029 "is_configured": true, 00:10:25.029 "data_offset": 0, 00:10:25.029 "data_size": 65536 00:10:25.029 }, 00:10:25.029 { 00:10:25.029 "name": "BaseBdev2", 00:10:25.029 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.029 "is_configured": false, 00:10:25.029 "data_offset": 0, 00:10:25.029 "data_size": 0 00:10:25.029 }, 00:10:25.029 { 00:10:25.029 "name": "BaseBdev3", 00:10:25.029 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.029 "is_configured": false, 00:10:25.029 "data_offset": 0, 00:10:25.029 "data_size": 0 00:10:25.029 } 00:10:25.029 ] 00:10:25.029 }' 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:25.029 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.287 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:25.287 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.287 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.287 [2024-11-20 15:58:23.519029] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:25.287 [2024-11-20 15:58:23.519064] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:10:25.287 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.287 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:25.287 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.287 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.287 [2024-11-20 15:58:23.527065] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:25.288 [2024-11-20 15:58:23.528570] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:25.288 [2024-11-20 15:58:23.528602] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:25.288 [2024-11-20 15:58:23.528610] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:25.288 [2024-11-20 15:58:23.528617] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:25.288 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.288 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:25.288 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:25.288 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:25.288 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:25.288 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:25.288 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:25.288 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:25.288 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:25.288 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:25.288 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:25.288 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:25.288 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:25.288 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.288 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:25.288 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.288 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.546 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.546 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:25.546 "name": "Existed_Raid", 00:10:25.546 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.546 "strip_size_kb": 0, 00:10:25.546 "state": "configuring", 00:10:25.546 "raid_level": "raid1", 00:10:25.546 "superblock": false, 00:10:25.546 "num_base_bdevs": 3, 00:10:25.546 "num_base_bdevs_discovered": 1, 00:10:25.546 "num_base_bdevs_operational": 3, 00:10:25.546 "base_bdevs_list": [ 00:10:25.546 { 00:10:25.546 "name": "BaseBdev1", 00:10:25.546 "uuid": "9dfe0b7c-5314-4793-a938-e488633e26a1", 00:10:25.546 "is_configured": true, 00:10:25.546 "data_offset": 0, 00:10:25.546 "data_size": 65536 00:10:25.546 }, 00:10:25.546 { 00:10:25.546 "name": "BaseBdev2", 00:10:25.546 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.546 "is_configured": false, 00:10:25.546 "data_offset": 0, 00:10:25.546 "data_size": 0 00:10:25.546 }, 00:10:25.546 { 00:10:25.546 "name": "BaseBdev3", 00:10:25.546 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.546 "is_configured": false, 00:10:25.546 "data_offset": 0, 00:10:25.546 "data_size": 0 00:10:25.546 } 00:10:25.546 ] 00:10:25.546 }' 00:10:25.546 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:25.546 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.804 [2024-11-20 15:58:23.841105] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:25.804 BaseBdev2 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.804 [ 00:10:25.804 { 00:10:25.804 "name": "BaseBdev2", 00:10:25.804 "aliases": [ 00:10:25.804 "6b330ed2-6e6e-4fc2-a2af-3a4747434c4e" 00:10:25.804 ], 00:10:25.804 "product_name": "Malloc disk", 00:10:25.804 "block_size": 512, 00:10:25.804 "num_blocks": 65536, 00:10:25.804 "uuid": "6b330ed2-6e6e-4fc2-a2af-3a4747434c4e", 00:10:25.804 "assigned_rate_limits": { 00:10:25.804 "rw_ios_per_sec": 0, 00:10:25.804 "rw_mbytes_per_sec": 0, 00:10:25.804 "r_mbytes_per_sec": 0, 00:10:25.804 "w_mbytes_per_sec": 0 00:10:25.804 }, 00:10:25.804 "claimed": true, 00:10:25.804 "claim_type": "exclusive_write", 00:10:25.804 "zoned": false, 00:10:25.804 "supported_io_types": { 00:10:25.804 "read": true, 00:10:25.804 "write": true, 00:10:25.804 "unmap": true, 00:10:25.804 "flush": true, 00:10:25.804 "reset": true, 00:10:25.804 "nvme_admin": false, 00:10:25.804 "nvme_io": false, 00:10:25.804 "nvme_io_md": false, 00:10:25.804 "write_zeroes": true, 00:10:25.804 "zcopy": true, 00:10:25.804 "get_zone_info": false, 00:10:25.804 "zone_management": false, 00:10:25.804 "zone_append": false, 00:10:25.804 "compare": false, 00:10:25.804 "compare_and_write": false, 00:10:25.804 "abort": true, 00:10:25.804 "seek_hole": false, 00:10:25.804 "seek_data": false, 00:10:25.804 "copy": true, 00:10:25.804 "nvme_iov_md": false 00:10:25.804 }, 00:10:25.804 "memory_domains": [ 00:10:25.804 { 00:10:25.804 "dma_device_id": "system", 00:10:25.804 "dma_device_type": 1 00:10:25.804 }, 00:10:25.804 { 00:10:25.804 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:25.804 "dma_device_type": 2 00:10:25.804 } 00:10:25.804 ], 00:10:25.804 "driver_specific": {} 00:10:25.804 } 00:10:25.804 ] 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:25.804 "name": "Existed_Raid", 00:10:25.804 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.804 "strip_size_kb": 0, 00:10:25.804 "state": "configuring", 00:10:25.804 "raid_level": "raid1", 00:10:25.804 "superblock": false, 00:10:25.804 "num_base_bdevs": 3, 00:10:25.804 "num_base_bdevs_discovered": 2, 00:10:25.804 "num_base_bdevs_operational": 3, 00:10:25.804 "base_bdevs_list": [ 00:10:25.804 { 00:10:25.804 "name": "BaseBdev1", 00:10:25.804 "uuid": "9dfe0b7c-5314-4793-a938-e488633e26a1", 00:10:25.804 "is_configured": true, 00:10:25.804 "data_offset": 0, 00:10:25.804 "data_size": 65536 00:10:25.804 }, 00:10:25.804 { 00:10:25.804 "name": "BaseBdev2", 00:10:25.804 "uuid": "6b330ed2-6e6e-4fc2-a2af-3a4747434c4e", 00:10:25.804 "is_configured": true, 00:10:25.804 "data_offset": 0, 00:10:25.804 "data_size": 65536 00:10:25.804 }, 00:10:25.804 { 00:10:25.804 "name": "BaseBdev3", 00:10:25.804 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.804 "is_configured": false, 00:10:25.804 "data_offset": 0, 00:10:25.804 "data_size": 0 00:10:25.804 } 00:10:25.804 ] 00:10:25.804 }' 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:25.804 15:58:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.063 [2024-11-20 15:58:24.207347] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:26.063 [2024-11-20 15:58:24.207505] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:10:26.063 [2024-11-20 15:58:24.207536] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:26.063 [2024-11-20 15:58:24.207828] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:10:26.063 [2024-11-20 15:58:24.208078] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:10:26.063 [2024-11-20 15:58:24.208092] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:10:26.063 [2024-11-20 15:58:24.208281] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:26.063 BaseBdev3 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.063 [ 00:10:26.063 { 00:10:26.063 "name": "BaseBdev3", 00:10:26.063 "aliases": [ 00:10:26.063 "eee26e94-1377-4c56-bc67-70492fd7dc69" 00:10:26.063 ], 00:10:26.063 "product_name": "Malloc disk", 00:10:26.063 "block_size": 512, 00:10:26.063 "num_blocks": 65536, 00:10:26.063 "uuid": "eee26e94-1377-4c56-bc67-70492fd7dc69", 00:10:26.063 "assigned_rate_limits": { 00:10:26.063 "rw_ios_per_sec": 0, 00:10:26.063 "rw_mbytes_per_sec": 0, 00:10:26.063 "r_mbytes_per_sec": 0, 00:10:26.063 "w_mbytes_per_sec": 0 00:10:26.063 }, 00:10:26.063 "claimed": true, 00:10:26.063 "claim_type": "exclusive_write", 00:10:26.063 "zoned": false, 00:10:26.063 "supported_io_types": { 00:10:26.063 "read": true, 00:10:26.063 "write": true, 00:10:26.063 "unmap": true, 00:10:26.063 "flush": true, 00:10:26.063 "reset": true, 00:10:26.063 "nvme_admin": false, 00:10:26.063 "nvme_io": false, 00:10:26.063 "nvme_io_md": false, 00:10:26.063 "write_zeroes": true, 00:10:26.063 "zcopy": true, 00:10:26.063 "get_zone_info": false, 00:10:26.063 "zone_management": false, 00:10:26.063 "zone_append": false, 00:10:26.063 "compare": false, 00:10:26.063 "compare_and_write": false, 00:10:26.063 "abort": true, 00:10:26.063 "seek_hole": false, 00:10:26.063 "seek_data": false, 00:10:26.063 "copy": true, 00:10:26.063 "nvme_iov_md": false 00:10:26.063 }, 00:10:26.063 "memory_domains": [ 00:10:26.063 { 00:10:26.063 "dma_device_id": "system", 00:10:26.063 "dma_device_type": 1 00:10:26.063 }, 00:10:26.063 { 00:10:26.063 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.063 "dma_device_type": 2 00:10:26.063 } 00:10:26.063 ], 00:10:26.063 "driver_specific": {} 00:10:26.063 } 00:10:26.063 ] 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:26.063 "name": "Existed_Raid", 00:10:26.063 "uuid": "6443ee16-a74a-45aa-8ed7-49a4c53e7560", 00:10:26.063 "strip_size_kb": 0, 00:10:26.063 "state": "online", 00:10:26.063 "raid_level": "raid1", 00:10:26.063 "superblock": false, 00:10:26.063 "num_base_bdevs": 3, 00:10:26.063 "num_base_bdevs_discovered": 3, 00:10:26.063 "num_base_bdevs_operational": 3, 00:10:26.063 "base_bdevs_list": [ 00:10:26.063 { 00:10:26.063 "name": "BaseBdev1", 00:10:26.063 "uuid": "9dfe0b7c-5314-4793-a938-e488633e26a1", 00:10:26.063 "is_configured": true, 00:10:26.063 "data_offset": 0, 00:10:26.063 "data_size": 65536 00:10:26.063 }, 00:10:26.063 { 00:10:26.063 "name": "BaseBdev2", 00:10:26.063 "uuid": "6b330ed2-6e6e-4fc2-a2af-3a4747434c4e", 00:10:26.063 "is_configured": true, 00:10:26.063 "data_offset": 0, 00:10:26.063 "data_size": 65536 00:10:26.063 }, 00:10:26.063 { 00:10:26.063 "name": "BaseBdev3", 00:10:26.063 "uuid": "eee26e94-1377-4c56-bc67-70492fd7dc69", 00:10:26.063 "is_configured": true, 00:10:26.063 "data_offset": 0, 00:10:26.063 "data_size": 65536 00:10:26.063 } 00:10:26.063 ] 00:10:26.063 }' 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:26.063 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.321 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:26.321 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:26.321 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:26.321 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:26.321 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:26.321 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:26.321 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:26.321 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:26.321 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.321 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.321 [2024-11-20 15:58:24.547737] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:26.321 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.321 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:26.321 "name": "Existed_Raid", 00:10:26.321 "aliases": [ 00:10:26.321 "6443ee16-a74a-45aa-8ed7-49a4c53e7560" 00:10:26.321 ], 00:10:26.321 "product_name": "Raid Volume", 00:10:26.321 "block_size": 512, 00:10:26.321 "num_blocks": 65536, 00:10:26.321 "uuid": "6443ee16-a74a-45aa-8ed7-49a4c53e7560", 00:10:26.321 "assigned_rate_limits": { 00:10:26.321 "rw_ios_per_sec": 0, 00:10:26.321 "rw_mbytes_per_sec": 0, 00:10:26.321 "r_mbytes_per_sec": 0, 00:10:26.321 "w_mbytes_per_sec": 0 00:10:26.321 }, 00:10:26.321 "claimed": false, 00:10:26.321 "zoned": false, 00:10:26.321 "supported_io_types": { 00:10:26.321 "read": true, 00:10:26.321 "write": true, 00:10:26.321 "unmap": false, 00:10:26.321 "flush": false, 00:10:26.321 "reset": true, 00:10:26.321 "nvme_admin": false, 00:10:26.321 "nvme_io": false, 00:10:26.321 "nvme_io_md": false, 00:10:26.321 "write_zeroes": true, 00:10:26.321 "zcopy": false, 00:10:26.321 "get_zone_info": false, 00:10:26.321 "zone_management": false, 00:10:26.321 "zone_append": false, 00:10:26.321 "compare": false, 00:10:26.321 "compare_and_write": false, 00:10:26.321 "abort": false, 00:10:26.321 "seek_hole": false, 00:10:26.321 "seek_data": false, 00:10:26.321 "copy": false, 00:10:26.321 "nvme_iov_md": false 00:10:26.321 }, 00:10:26.321 "memory_domains": [ 00:10:26.321 { 00:10:26.321 "dma_device_id": "system", 00:10:26.321 "dma_device_type": 1 00:10:26.321 }, 00:10:26.321 { 00:10:26.321 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.321 "dma_device_type": 2 00:10:26.321 }, 00:10:26.321 { 00:10:26.321 "dma_device_id": "system", 00:10:26.321 "dma_device_type": 1 00:10:26.321 }, 00:10:26.321 { 00:10:26.321 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.321 "dma_device_type": 2 00:10:26.321 }, 00:10:26.322 { 00:10:26.322 "dma_device_id": "system", 00:10:26.322 "dma_device_type": 1 00:10:26.322 }, 00:10:26.322 { 00:10:26.322 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.322 "dma_device_type": 2 00:10:26.322 } 00:10:26.322 ], 00:10:26.322 "driver_specific": { 00:10:26.322 "raid": { 00:10:26.322 "uuid": "6443ee16-a74a-45aa-8ed7-49a4c53e7560", 00:10:26.322 "strip_size_kb": 0, 00:10:26.322 "state": "online", 00:10:26.322 "raid_level": "raid1", 00:10:26.322 "superblock": false, 00:10:26.322 "num_base_bdevs": 3, 00:10:26.322 "num_base_bdevs_discovered": 3, 00:10:26.322 "num_base_bdevs_operational": 3, 00:10:26.322 "base_bdevs_list": [ 00:10:26.322 { 00:10:26.322 "name": "BaseBdev1", 00:10:26.322 "uuid": "9dfe0b7c-5314-4793-a938-e488633e26a1", 00:10:26.322 "is_configured": true, 00:10:26.322 "data_offset": 0, 00:10:26.322 "data_size": 65536 00:10:26.322 }, 00:10:26.322 { 00:10:26.322 "name": "BaseBdev2", 00:10:26.322 "uuid": "6b330ed2-6e6e-4fc2-a2af-3a4747434c4e", 00:10:26.322 "is_configured": true, 00:10:26.322 "data_offset": 0, 00:10:26.322 "data_size": 65536 00:10:26.322 }, 00:10:26.322 { 00:10:26.322 "name": "BaseBdev3", 00:10:26.322 "uuid": "eee26e94-1377-4c56-bc67-70492fd7dc69", 00:10:26.322 "is_configured": true, 00:10:26.322 "data_offset": 0, 00:10:26.322 "data_size": 65536 00:10:26.322 } 00:10:26.322 ] 00:10:26.322 } 00:10:26.322 } 00:10:26.322 }' 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:26.580 BaseBdev2 00:10:26.580 BaseBdev3' 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.580 [2024-11-20 15:58:24.727531] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:26.580 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:26.581 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:26.581 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:26.581 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:26.581 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:26.581 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:26.581 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.581 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:26.581 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.581 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.581 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.581 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:26.581 "name": "Existed_Raid", 00:10:26.581 "uuid": "6443ee16-a74a-45aa-8ed7-49a4c53e7560", 00:10:26.581 "strip_size_kb": 0, 00:10:26.581 "state": "online", 00:10:26.581 "raid_level": "raid1", 00:10:26.581 "superblock": false, 00:10:26.581 "num_base_bdevs": 3, 00:10:26.581 "num_base_bdevs_discovered": 2, 00:10:26.581 "num_base_bdevs_operational": 2, 00:10:26.581 "base_bdevs_list": [ 00:10:26.581 { 00:10:26.581 "name": null, 00:10:26.581 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:26.581 "is_configured": false, 00:10:26.581 "data_offset": 0, 00:10:26.581 "data_size": 65536 00:10:26.581 }, 00:10:26.581 { 00:10:26.581 "name": "BaseBdev2", 00:10:26.581 "uuid": "6b330ed2-6e6e-4fc2-a2af-3a4747434c4e", 00:10:26.581 "is_configured": true, 00:10:26.581 "data_offset": 0, 00:10:26.581 "data_size": 65536 00:10:26.581 }, 00:10:26.581 { 00:10:26.581 "name": "BaseBdev3", 00:10:26.581 "uuid": "eee26e94-1377-4c56-bc67-70492fd7dc69", 00:10:26.581 "is_configured": true, 00:10:26.581 "data_offset": 0, 00:10:26.581 "data_size": 65536 00:10:26.581 } 00:10:26.581 ] 00:10:26.581 }' 00:10:26.581 15:58:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:26.581 15:58:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.147 [2024-11-20 15:58:25.129573] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.147 [2024-11-20 15:58:25.215969] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:27.147 [2024-11-20 15:58:25.216040] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:27.147 [2024-11-20 15:58:25.262457] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:27.147 [2024-11-20 15:58:25.262606] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:27.147 [2024-11-20 15:58:25.262743] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.147 BaseBdev2 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.147 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.148 [ 00:10:27.148 { 00:10:27.148 "name": "BaseBdev2", 00:10:27.148 "aliases": [ 00:10:27.148 "44f1efe4-33b7-40cc-b00b-231d7354d280" 00:10:27.148 ], 00:10:27.148 "product_name": "Malloc disk", 00:10:27.148 "block_size": 512, 00:10:27.148 "num_blocks": 65536, 00:10:27.148 "uuid": "44f1efe4-33b7-40cc-b00b-231d7354d280", 00:10:27.148 "assigned_rate_limits": { 00:10:27.148 "rw_ios_per_sec": 0, 00:10:27.148 "rw_mbytes_per_sec": 0, 00:10:27.148 "r_mbytes_per_sec": 0, 00:10:27.148 "w_mbytes_per_sec": 0 00:10:27.148 }, 00:10:27.148 "claimed": false, 00:10:27.148 "zoned": false, 00:10:27.148 "supported_io_types": { 00:10:27.148 "read": true, 00:10:27.148 "write": true, 00:10:27.148 "unmap": true, 00:10:27.148 "flush": true, 00:10:27.148 "reset": true, 00:10:27.148 "nvme_admin": false, 00:10:27.148 "nvme_io": false, 00:10:27.148 "nvme_io_md": false, 00:10:27.148 "write_zeroes": true, 00:10:27.148 "zcopy": true, 00:10:27.148 "get_zone_info": false, 00:10:27.148 "zone_management": false, 00:10:27.148 "zone_append": false, 00:10:27.148 "compare": false, 00:10:27.148 "compare_and_write": false, 00:10:27.148 "abort": true, 00:10:27.148 "seek_hole": false, 00:10:27.148 "seek_data": false, 00:10:27.148 "copy": true, 00:10:27.148 "nvme_iov_md": false 00:10:27.148 }, 00:10:27.148 "memory_domains": [ 00:10:27.148 { 00:10:27.148 "dma_device_id": "system", 00:10:27.148 "dma_device_type": 1 00:10:27.148 }, 00:10:27.148 { 00:10:27.148 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:27.148 "dma_device_type": 2 00:10:27.148 } 00:10:27.148 ], 00:10:27.148 "driver_specific": {} 00:10:27.148 } 00:10:27.148 ] 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.148 BaseBdev3 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.148 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.148 [ 00:10:27.148 { 00:10:27.148 "name": "BaseBdev3", 00:10:27.148 "aliases": [ 00:10:27.148 "11c1f54e-2c56-453a-b803-3260bae925e6" 00:10:27.148 ], 00:10:27.148 "product_name": "Malloc disk", 00:10:27.148 "block_size": 512, 00:10:27.148 "num_blocks": 65536, 00:10:27.148 "uuid": "11c1f54e-2c56-453a-b803-3260bae925e6", 00:10:27.148 "assigned_rate_limits": { 00:10:27.148 "rw_ios_per_sec": 0, 00:10:27.148 "rw_mbytes_per_sec": 0, 00:10:27.148 "r_mbytes_per_sec": 0, 00:10:27.148 "w_mbytes_per_sec": 0 00:10:27.148 }, 00:10:27.148 "claimed": false, 00:10:27.148 "zoned": false, 00:10:27.148 "supported_io_types": { 00:10:27.148 "read": true, 00:10:27.148 "write": true, 00:10:27.148 "unmap": true, 00:10:27.148 "flush": true, 00:10:27.148 "reset": true, 00:10:27.148 "nvme_admin": false, 00:10:27.148 "nvme_io": false, 00:10:27.148 "nvme_io_md": false, 00:10:27.148 "write_zeroes": true, 00:10:27.148 "zcopy": true, 00:10:27.148 "get_zone_info": false, 00:10:27.148 "zone_management": false, 00:10:27.148 "zone_append": false, 00:10:27.148 "compare": false, 00:10:27.148 "compare_and_write": false, 00:10:27.148 "abort": true, 00:10:27.148 "seek_hole": false, 00:10:27.148 "seek_data": false, 00:10:27.406 "copy": true, 00:10:27.406 "nvme_iov_md": false 00:10:27.406 }, 00:10:27.406 "memory_domains": [ 00:10:27.406 { 00:10:27.406 "dma_device_id": "system", 00:10:27.406 "dma_device_type": 1 00:10:27.406 }, 00:10:27.406 { 00:10:27.406 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:27.406 "dma_device_type": 2 00:10:27.406 } 00:10:27.406 ], 00:10:27.406 "driver_specific": {} 00:10:27.406 } 00:10:27.406 ] 00:10:27.406 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.406 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:27.406 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:27.406 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:27.406 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:27.406 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.406 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.406 [2024-11-20 15:58:25.402267] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:27.406 [2024-11-20 15:58:25.402395] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:27.406 [2024-11-20 15:58:25.402464] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:27.406 [2024-11-20 15:58:25.404052] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:27.406 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.406 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:27.406 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:27.406 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:27.406 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:27.406 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:27.406 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:27.406 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:27.406 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:27.406 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:27.407 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:27.407 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.407 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.407 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.407 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:27.407 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.407 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:27.407 "name": "Existed_Raid", 00:10:27.407 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:27.407 "strip_size_kb": 0, 00:10:27.407 "state": "configuring", 00:10:27.407 "raid_level": "raid1", 00:10:27.407 "superblock": false, 00:10:27.407 "num_base_bdevs": 3, 00:10:27.407 "num_base_bdevs_discovered": 2, 00:10:27.407 "num_base_bdevs_operational": 3, 00:10:27.407 "base_bdevs_list": [ 00:10:27.407 { 00:10:27.407 "name": "BaseBdev1", 00:10:27.407 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:27.407 "is_configured": false, 00:10:27.407 "data_offset": 0, 00:10:27.407 "data_size": 0 00:10:27.407 }, 00:10:27.407 { 00:10:27.407 "name": "BaseBdev2", 00:10:27.407 "uuid": "44f1efe4-33b7-40cc-b00b-231d7354d280", 00:10:27.407 "is_configured": true, 00:10:27.407 "data_offset": 0, 00:10:27.407 "data_size": 65536 00:10:27.407 }, 00:10:27.407 { 00:10:27.407 "name": "BaseBdev3", 00:10:27.407 "uuid": "11c1f54e-2c56-453a-b803-3260bae925e6", 00:10:27.407 "is_configured": true, 00:10:27.407 "data_offset": 0, 00:10:27.407 "data_size": 65536 00:10:27.407 } 00:10:27.407 ] 00:10:27.407 }' 00:10:27.407 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:27.407 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.664 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:27.664 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.664 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.664 [2024-11-20 15:58:25.734362] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:27.664 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.664 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:27.665 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:27.665 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:27.665 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:27.665 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:27.665 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:27.665 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:27.665 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:27.665 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:27.665 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:27.665 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.665 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:27.665 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.665 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.665 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.665 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:27.665 "name": "Existed_Raid", 00:10:27.665 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:27.665 "strip_size_kb": 0, 00:10:27.665 "state": "configuring", 00:10:27.665 "raid_level": "raid1", 00:10:27.665 "superblock": false, 00:10:27.665 "num_base_bdevs": 3, 00:10:27.665 "num_base_bdevs_discovered": 1, 00:10:27.665 "num_base_bdevs_operational": 3, 00:10:27.665 "base_bdevs_list": [ 00:10:27.665 { 00:10:27.665 "name": "BaseBdev1", 00:10:27.665 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:27.665 "is_configured": false, 00:10:27.665 "data_offset": 0, 00:10:27.665 "data_size": 0 00:10:27.665 }, 00:10:27.665 { 00:10:27.665 "name": null, 00:10:27.665 "uuid": "44f1efe4-33b7-40cc-b00b-231d7354d280", 00:10:27.665 "is_configured": false, 00:10:27.665 "data_offset": 0, 00:10:27.665 "data_size": 65536 00:10:27.665 }, 00:10:27.665 { 00:10:27.665 "name": "BaseBdev3", 00:10:27.665 "uuid": "11c1f54e-2c56-453a-b803-3260bae925e6", 00:10:27.665 "is_configured": true, 00:10:27.665 "data_offset": 0, 00:10:27.665 "data_size": 65536 00:10:27.665 } 00:10:27.665 ] 00:10:27.665 }' 00:10:27.665 15:58:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:27.665 15:58:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.924 [2024-11-20 15:58:26.112695] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:27.924 BaseBdev1 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.924 [ 00:10:27.924 { 00:10:27.924 "name": "BaseBdev1", 00:10:27.924 "aliases": [ 00:10:27.924 "d8cede6b-72b8-4aa6-a187-972969d27f64" 00:10:27.924 ], 00:10:27.924 "product_name": "Malloc disk", 00:10:27.924 "block_size": 512, 00:10:27.924 "num_blocks": 65536, 00:10:27.924 "uuid": "d8cede6b-72b8-4aa6-a187-972969d27f64", 00:10:27.924 "assigned_rate_limits": { 00:10:27.924 "rw_ios_per_sec": 0, 00:10:27.924 "rw_mbytes_per_sec": 0, 00:10:27.924 "r_mbytes_per_sec": 0, 00:10:27.924 "w_mbytes_per_sec": 0 00:10:27.924 }, 00:10:27.924 "claimed": true, 00:10:27.924 "claim_type": "exclusive_write", 00:10:27.924 "zoned": false, 00:10:27.924 "supported_io_types": { 00:10:27.924 "read": true, 00:10:27.924 "write": true, 00:10:27.924 "unmap": true, 00:10:27.924 "flush": true, 00:10:27.924 "reset": true, 00:10:27.924 "nvme_admin": false, 00:10:27.924 "nvme_io": false, 00:10:27.924 "nvme_io_md": false, 00:10:27.924 "write_zeroes": true, 00:10:27.924 "zcopy": true, 00:10:27.924 "get_zone_info": false, 00:10:27.924 "zone_management": false, 00:10:27.924 "zone_append": false, 00:10:27.924 "compare": false, 00:10:27.924 "compare_and_write": false, 00:10:27.924 "abort": true, 00:10:27.924 "seek_hole": false, 00:10:27.924 "seek_data": false, 00:10:27.924 "copy": true, 00:10:27.924 "nvme_iov_md": false 00:10:27.924 }, 00:10:27.924 "memory_domains": [ 00:10:27.924 { 00:10:27.924 "dma_device_id": "system", 00:10:27.924 "dma_device_type": 1 00:10:27.924 }, 00:10:27.924 { 00:10:27.924 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:27.924 "dma_device_type": 2 00:10:27.924 } 00:10:27.924 ], 00:10:27.924 "driver_specific": {} 00:10:27.924 } 00:10:27.924 ] 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:27.924 "name": "Existed_Raid", 00:10:27.924 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:27.924 "strip_size_kb": 0, 00:10:27.924 "state": "configuring", 00:10:27.924 "raid_level": "raid1", 00:10:27.924 "superblock": false, 00:10:27.924 "num_base_bdevs": 3, 00:10:27.924 "num_base_bdevs_discovered": 2, 00:10:27.924 "num_base_bdevs_operational": 3, 00:10:27.924 "base_bdevs_list": [ 00:10:27.924 { 00:10:27.924 "name": "BaseBdev1", 00:10:27.924 "uuid": "d8cede6b-72b8-4aa6-a187-972969d27f64", 00:10:27.924 "is_configured": true, 00:10:27.924 "data_offset": 0, 00:10:27.924 "data_size": 65536 00:10:27.924 }, 00:10:27.924 { 00:10:27.924 "name": null, 00:10:27.924 "uuid": "44f1efe4-33b7-40cc-b00b-231d7354d280", 00:10:27.924 "is_configured": false, 00:10:27.924 "data_offset": 0, 00:10:27.924 "data_size": 65536 00:10:27.924 }, 00:10:27.924 { 00:10:27.924 "name": "BaseBdev3", 00:10:27.924 "uuid": "11c1f54e-2c56-453a-b803-3260bae925e6", 00:10:27.924 "is_configured": true, 00:10:27.924 "data_offset": 0, 00:10:27.924 "data_size": 65536 00:10:27.924 } 00:10:27.924 ] 00:10:27.924 }' 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:27.924 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.489 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.489 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.489 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.489 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:28.489 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.489 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:28.489 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.490 [2024-11-20 15:58:26.492802] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:28.490 "name": "Existed_Raid", 00:10:28.490 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:28.490 "strip_size_kb": 0, 00:10:28.490 "state": "configuring", 00:10:28.490 "raid_level": "raid1", 00:10:28.490 "superblock": false, 00:10:28.490 "num_base_bdevs": 3, 00:10:28.490 "num_base_bdevs_discovered": 1, 00:10:28.490 "num_base_bdevs_operational": 3, 00:10:28.490 "base_bdevs_list": [ 00:10:28.490 { 00:10:28.490 "name": "BaseBdev1", 00:10:28.490 "uuid": "d8cede6b-72b8-4aa6-a187-972969d27f64", 00:10:28.490 "is_configured": true, 00:10:28.490 "data_offset": 0, 00:10:28.490 "data_size": 65536 00:10:28.490 }, 00:10:28.490 { 00:10:28.490 "name": null, 00:10:28.490 "uuid": "44f1efe4-33b7-40cc-b00b-231d7354d280", 00:10:28.490 "is_configured": false, 00:10:28.490 "data_offset": 0, 00:10:28.490 "data_size": 65536 00:10:28.490 }, 00:10:28.490 { 00:10:28.490 "name": null, 00:10:28.490 "uuid": "11c1f54e-2c56-453a-b803-3260bae925e6", 00:10:28.490 "is_configured": false, 00:10:28.490 "data_offset": 0, 00:10:28.490 "data_size": 65536 00:10:28.490 } 00:10:28.490 ] 00:10:28.490 }' 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:28.490 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.747 [2024-11-20 15:58:26.856914] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:28.747 "name": "Existed_Raid", 00:10:28.747 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:28.747 "strip_size_kb": 0, 00:10:28.747 "state": "configuring", 00:10:28.747 "raid_level": "raid1", 00:10:28.747 "superblock": false, 00:10:28.747 "num_base_bdevs": 3, 00:10:28.747 "num_base_bdevs_discovered": 2, 00:10:28.747 "num_base_bdevs_operational": 3, 00:10:28.747 "base_bdevs_list": [ 00:10:28.747 { 00:10:28.747 "name": "BaseBdev1", 00:10:28.747 "uuid": "d8cede6b-72b8-4aa6-a187-972969d27f64", 00:10:28.747 "is_configured": true, 00:10:28.747 "data_offset": 0, 00:10:28.747 "data_size": 65536 00:10:28.747 }, 00:10:28.747 { 00:10:28.747 "name": null, 00:10:28.747 "uuid": "44f1efe4-33b7-40cc-b00b-231d7354d280", 00:10:28.747 "is_configured": false, 00:10:28.747 "data_offset": 0, 00:10:28.747 "data_size": 65536 00:10:28.747 }, 00:10:28.747 { 00:10:28.747 "name": "BaseBdev3", 00:10:28.747 "uuid": "11c1f54e-2c56-453a-b803-3260bae925e6", 00:10:28.747 "is_configured": true, 00:10:28.747 "data_offset": 0, 00:10:28.747 "data_size": 65536 00:10:28.747 } 00:10:28.747 ] 00:10:28.747 }' 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:28.747 15:58:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.007 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.007 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.007 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.007 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:29.007 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.007 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:29.007 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:29.007 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.007 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.007 [2024-11-20 15:58:27.236987] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:29.265 "name": "Existed_Raid", 00:10:29.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:29.265 "strip_size_kb": 0, 00:10:29.265 "state": "configuring", 00:10:29.265 "raid_level": "raid1", 00:10:29.265 "superblock": false, 00:10:29.265 "num_base_bdevs": 3, 00:10:29.265 "num_base_bdevs_discovered": 1, 00:10:29.265 "num_base_bdevs_operational": 3, 00:10:29.265 "base_bdevs_list": [ 00:10:29.265 { 00:10:29.265 "name": null, 00:10:29.265 "uuid": "d8cede6b-72b8-4aa6-a187-972969d27f64", 00:10:29.265 "is_configured": false, 00:10:29.265 "data_offset": 0, 00:10:29.265 "data_size": 65536 00:10:29.265 }, 00:10:29.265 { 00:10:29.265 "name": null, 00:10:29.265 "uuid": "44f1efe4-33b7-40cc-b00b-231d7354d280", 00:10:29.265 "is_configured": false, 00:10:29.265 "data_offset": 0, 00:10:29.265 "data_size": 65536 00:10:29.265 }, 00:10:29.265 { 00:10:29.265 "name": "BaseBdev3", 00:10:29.265 "uuid": "11c1f54e-2c56-453a-b803-3260bae925e6", 00:10:29.265 "is_configured": true, 00:10:29.265 "data_offset": 0, 00:10:29.265 "data_size": 65536 00:10:29.265 } 00:10:29.265 ] 00:10:29.265 }' 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:29.265 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.523 [2024-11-20 15:58:27.635742] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:29.523 "name": "Existed_Raid", 00:10:29.523 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:29.523 "strip_size_kb": 0, 00:10:29.523 "state": "configuring", 00:10:29.523 "raid_level": "raid1", 00:10:29.523 "superblock": false, 00:10:29.523 "num_base_bdevs": 3, 00:10:29.523 "num_base_bdevs_discovered": 2, 00:10:29.523 "num_base_bdevs_operational": 3, 00:10:29.523 "base_bdevs_list": [ 00:10:29.523 { 00:10:29.523 "name": null, 00:10:29.523 "uuid": "d8cede6b-72b8-4aa6-a187-972969d27f64", 00:10:29.523 "is_configured": false, 00:10:29.523 "data_offset": 0, 00:10:29.523 "data_size": 65536 00:10:29.523 }, 00:10:29.523 { 00:10:29.523 "name": "BaseBdev2", 00:10:29.523 "uuid": "44f1efe4-33b7-40cc-b00b-231d7354d280", 00:10:29.523 "is_configured": true, 00:10:29.523 "data_offset": 0, 00:10:29.523 "data_size": 65536 00:10:29.523 }, 00:10:29.523 { 00:10:29.523 "name": "BaseBdev3", 00:10:29.523 "uuid": "11c1f54e-2c56-453a-b803-3260bae925e6", 00:10:29.523 "is_configured": true, 00:10:29.523 "data_offset": 0, 00:10:29.523 "data_size": 65536 00:10:29.523 } 00:10:29.523 ] 00:10:29.523 }' 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:29.523 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.782 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.782 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.782 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.782 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:29.782 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.782 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:29.782 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.782 15:58:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:29.782 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.782 15:58:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.782 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.782 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u d8cede6b-72b8-4aa6-a187-972969d27f64 00:10:29.782 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.782 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.040 [2024-11-20 15:58:28.046068] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:30.040 [2024-11-20 15:58:28.046106] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:10:30.040 [2024-11-20 15:58:28.046112] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:30.040 [2024-11-20 15:58:28.046307] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:10:30.040 [2024-11-20 15:58:28.046409] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:10:30.040 [2024-11-20 15:58:28.046417] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:10:30.040 [2024-11-20 15:58:28.046592] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:30.040 NewBaseBdev 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.040 [ 00:10:30.040 { 00:10:30.040 "name": "NewBaseBdev", 00:10:30.040 "aliases": [ 00:10:30.040 "d8cede6b-72b8-4aa6-a187-972969d27f64" 00:10:30.040 ], 00:10:30.040 "product_name": "Malloc disk", 00:10:30.040 "block_size": 512, 00:10:30.040 "num_blocks": 65536, 00:10:30.040 "uuid": "d8cede6b-72b8-4aa6-a187-972969d27f64", 00:10:30.040 "assigned_rate_limits": { 00:10:30.040 "rw_ios_per_sec": 0, 00:10:30.040 "rw_mbytes_per_sec": 0, 00:10:30.040 "r_mbytes_per_sec": 0, 00:10:30.040 "w_mbytes_per_sec": 0 00:10:30.040 }, 00:10:30.040 "claimed": true, 00:10:30.040 "claim_type": "exclusive_write", 00:10:30.040 "zoned": false, 00:10:30.040 "supported_io_types": { 00:10:30.040 "read": true, 00:10:30.040 "write": true, 00:10:30.040 "unmap": true, 00:10:30.040 "flush": true, 00:10:30.040 "reset": true, 00:10:30.040 "nvme_admin": false, 00:10:30.040 "nvme_io": false, 00:10:30.040 "nvme_io_md": false, 00:10:30.040 "write_zeroes": true, 00:10:30.040 "zcopy": true, 00:10:30.040 "get_zone_info": false, 00:10:30.040 "zone_management": false, 00:10:30.040 "zone_append": false, 00:10:30.040 "compare": false, 00:10:30.040 "compare_and_write": false, 00:10:30.040 "abort": true, 00:10:30.040 "seek_hole": false, 00:10:30.040 "seek_data": false, 00:10:30.040 "copy": true, 00:10:30.040 "nvme_iov_md": false 00:10:30.040 }, 00:10:30.040 "memory_domains": [ 00:10:30.040 { 00:10:30.040 "dma_device_id": "system", 00:10:30.040 "dma_device_type": 1 00:10:30.040 }, 00:10:30.040 { 00:10:30.040 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.040 "dma_device_type": 2 00:10:30.040 } 00:10:30.040 ], 00:10:30.040 "driver_specific": {} 00:10:30.040 } 00:10:30.040 ] 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:30.040 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:30.041 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:30.041 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:30.041 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:30.041 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:30.041 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:30.041 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.041 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.041 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.041 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:30.041 "name": "Existed_Raid", 00:10:30.041 "uuid": "955eb5b9-993c-4e7a-9b4e-8c29c1c7188b", 00:10:30.041 "strip_size_kb": 0, 00:10:30.041 "state": "online", 00:10:30.041 "raid_level": "raid1", 00:10:30.041 "superblock": false, 00:10:30.041 "num_base_bdevs": 3, 00:10:30.041 "num_base_bdevs_discovered": 3, 00:10:30.041 "num_base_bdevs_operational": 3, 00:10:30.041 "base_bdevs_list": [ 00:10:30.041 { 00:10:30.041 "name": "NewBaseBdev", 00:10:30.041 "uuid": "d8cede6b-72b8-4aa6-a187-972969d27f64", 00:10:30.041 "is_configured": true, 00:10:30.041 "data_offset": 0, 00:10:30.041 "data_size": 65536 00:10:30.041 }, 00:10:30.041 { 00:10:30.041 "name": "BaseBdev2", 00:10:30.041 "uuid": "44f1efe4-33b7-40cc-b00b-231d7354d280", 00:10:30.041 "is_configured": true, 00:10:30.041 "data_offset": 0, 00:10:30.041 "data_size": 65536 00:10:30.041 }, 00:10:30.041 { 00:10:30.041 "name": "BaseBdev3", 00:10:30.041 "uuid": "11c1f54e-2c56-453a-b803-3260bae925e6", 00:10:30.041 "is_configured": true, 00:10:30.041 "data_offset": 0, 00:10:30.041 "data_size": 65536 00:10:30.041 } 00:10:30.041 ] 00:10:30.041 }' 00:10:30.041 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:30.041 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.298 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:30.298 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.299 [2024-11-20 15:58:28.374436] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:30.299 "name": "Existed_Raid", 00:10:30.299 "aliases": [ 00:10:30.299 "955eb5b9-993c-4e7a-9b4e-8c29c1c7188b" 00:10:30.299 ], 00:10:30.299 "product_name": "Raid Volume", 00:10:30.299 "block_size": 512, 00:10:30.299 "num_blocks": 65536, 00:10:30.299 "uuid": "955eb5b9-993c-4e7a-9b4e-8c29c1c7188b", 00:10:30.299 "assigned_rate_limits": { 00:10:30.299 "rw_ios_per_sec": 0, 00:10:30.299 "rw_mbytes_per_sec": 0, 00:10:30.299 "r_mbytes_per_sec": 0, 00:10:30.299 "w_mbytes_per_sec": 0 00:10:30.299 }, 00:10:30.299 "claimed": false, 00:10:30.299 "zoned": false, 00:10:30.299 "supported_io_types": { 00:10:30.299 "read": true, 00:10:30.299 "write": true, 00:10:30.299 "unmap": false, 00:10:30.299 "flush": false, 00:10:30.299 "reset": true, 00:10:30.299 "nvme_admin": false, 00:10:30.299 "nvme_io": false, 00:10:30.299 "nvme_io_md": false, 00:10:30.299 "write_zeroes": true, 00:10:30.299 "zcopy": false, 00:10:30.299 "get_zone_info": false, 00:10:30.299 "zone_management": false, 00:10:30.299 "zone_append": false, 00:10:30.299 "compare": false, 00:10:30.299 "compare_and_write": false, 00:10:30.299 "abort": false, 00:10:30.299 "seek_hole": false, 00:10:30.299 "seek_data": false, 00:10:30.299 "copy": false, 00:10:30.299 "nvme_iov_md": false 00:10:30.299 }, 00:10:30.299 "memory_domains": [ 00:10:30.299 { 00:10:30.299 "dma_device_id": "system", 00:10:30.299 "dma_device_type": 1 00:10:30.299 }, 00:10:30.299 { 00:10:30.299 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.299 "dma_device_type": 2 00:10:30.299 }, 00:10:30.299 { 00:10:30.299 "dma_device_id": "system", 00:10:30.299 "dma_device_type": 1 00:10:30.299 }, 00:10:30.299 { 00:10:30.299 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.299 "dma_device_type": 2 00:10:30.299 }, 00:10:30.299 { 00:10:30.299 "dma_device_id": "system", 00:10:30.299 "dma_device_type": 1 00:10:30.299 }, 00:10:30.299 { 00:10:30.299 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.299 "dma_device_type": 2 00:10:30.299 } 00:10:30.299 ], 00:10:30.299 "driver_specific": { 00:10:30.299 "raid": { 00:10:30.299 "uuid": "955eb5b9-993c-4e7a-9b4e-8c29c1c7188b", 00:10:30.299 "strip_size_kb": 0, 00:10:30.299 "state": "online", 00:10:30.299 "raid_level": "raid1", 00:10:30.299 "superblock": false, 00:10:30.299 "num_base_bdevs": 3, 00:10:30.299 "num_base_bdevs_discovered": 3, 00:10:30.299 "num_base_bdevs_operational": 3, 00:10:30.299 "base_bdevs_list": [ 00:10:30.299 { 00:10:30.299 "name": "NewBaseBdev", 00:10:30.299 "uuid": "d8cede6b-72b8-4aa6-a187-972969d27f64", 00:10:30.299 "is_configured": true, 00:10:30.299 "data_offset": 0, 00:10:30.299 "data_size": 65536 00:10:30.299 }, 00:10:30.299 { 00:10:30.299 "name": "BaseBdev2", 00:10:30.299 "uuid": "44f1efe4-33b7-40cc-b00b-231d7354d280", 00:10:30.299 "is_configured": true, 00:10:30.299 "data_offset": 0, 00:10:30.299 "data_size": 65536 00:10:30.299 }, 00:10:30.299 { 00:10:30.299 "name": "BaseBdev3", 00:10:30.299 "uuid": "11c1f54e-2c56-453a-b803-3260bae925e6", 00:10:30.299 "is_configured": true, 00:10:30.299 "data_offset": 0, 00:10:30.299 "data_size": 65536 00:10:30.299 } 00:10:30.299 ] 00:10:30.299 } 00:10:30.299 } 00:10:30.299 }' 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:30.299 BaseBdev2 00:10:30.299 BaseBdev3' 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.299 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.300 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.557 [2024-11-20 15:58:28.570208] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:30.557 [2024-11-20 15:58:28.570326] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:30.557 [2024-11-20 15:58:28.570394] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:30.557 [2024-11-20 15:58:28.570613] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:30.557 [2024-11-20 15:58:28.570621] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 65701 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 65701 ']' 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 65701 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 65701 00:10:30.557 killing process with pid 65701 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 65701' 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 65701 00:10:30.557 [2024-11-20 15:58:28.602419] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:30.557 15:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 65701 00:10:30.557 [2024-11-20 15:58:28.748623] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:10:31.123 00:10:31.123 real 0m7.389s 00:10:31.123 user 0m11.926s 00:10:31.123 sys 0m1.238s 00:10:31.123 ************************************ 00:10:31.123 END TEST raid_state_function_test 00:10:31.123 ************************************ 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.123 15:58:29 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 3 true 00:10:31.123 15:58:29 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:31.123 15:58:29 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:31.123 15:58:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:31.123 ************************************ 00:10:31.123 START TEST raid_state_function_test_sb 00:10:31.123 ************************************ 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 3 true 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=66290 00:10:31.123 Process raid pid: 66290 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 66290' 00:10:31.123 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 66290 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 66290 ']' 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.123 15:58:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:31.381 [2024-11-20 15:58:29.420766] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:10:31.381 [2024-11-20 15:58:29.420860] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:31.381 [2024-11-20 15:58:29.570458] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:31.638 [2024-11-20 15:58:29.655339] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:31.638 [2024-11-20 15:58:29.765643] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:31.638 [2024-11-20 15:58:29.765687] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.203 [2024-11-20 15:58:30.276720] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:32.203 [2024-11-20 15:58:30.276760] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:32.203 [2024-11-20 15:58:30.276768] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:32.203 [2024-11-20 15:58:30.276776] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:32.203 [2024-11-20 15:58:30.276781] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:32.203 [2024-11-20 15:58:30.276788] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:32.203 "name": "Existed_Raid", 00:10:32.203 "uuid": "9a78e1af-95af-41dd-851b-4d5b876902b8", 00:10:32.203 "strip_size_kb": 0, 00:10:32.203 "state": "configuring", 00:10:32.203 "raid_level": "raid1", 00:10:32.203 "superblock": true, 00:10:32.203 "num_base_bdevs": 3, 00:10:32.203 "num_base_bdevs_discovered": 0, 00:10:32.203 "num_base_bdevs_operational": 3, 00:10:32.203 "base_bdevs_list": [ 00:10:32.203 { 00:10:32.203 "name": "BaseBdev1", 00:10:32.203 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.203 "is_configured": false, 00:10:32.203 "data_offset": 0, 00:10:32.203 "data_size": 0 00:10:32.203 }, 00:10:32.203 { 00:10:32.203 "name": "BaseBdev2", 00:10:32.203 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.203 "is_configured": false, 00:10:32.203 "data_offset": 0, 00:10:32.203 "data_size": 0 00:10:32.203 }, 00:10:32.203 { 00:10:32.203 "name": "BaseBdev3", 00:10:32.203 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.203 "is_configured": false, 00:10:32.203 "data_offset": 0, 00:10:32.203 "data_size": 0 00:10:32.203 } 00:10:32.203 ] 00:10:32.203 }' 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:32.203 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.461 [2024-11-20 15:58:30.604724] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:32.461 [2024-11-20 15:58:30.604751] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.461 [2024-11-20 15:58:30.612722] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:32.461 [2024-11-20 15:58:30.612754] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:32.461 [2024-11-20 15:58:30.612761] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:32.461 [2024-11-20 15:58:30.612768] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:32.461 [2024-11-20 15:58:30.612773] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:32.461 [2024-11-20 15:58:30.612779] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.461 [2024-11-20 15:58:30.640516] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:32.461 BaseBdev1 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.461 [ 00:10:32.461 { 00:10:32.461 "name": "BaseBdev1", 00:10:32.461 "aliases": [ 00:10:32.461 "598364e3-d476-441c-99da-b8434c2142d0" 00:10:32.461 ], 00:10:32.461 "product_name": "Malloc disk", 00:10:32.461 "block_size": 512, 00:10:32.461 "num_blocks": 65536, 00:10:32.461 "uuid": "598364e3-d476-441c-99da-b8434c2142d0", 00:10:32.461 "assigned_rate_limits": { 00:10:32.461 "rw_ios_per_sec": 0, 00:10:32.461 "rw_mbytes_per_sec": 0, 00:10:32.461 "r_mbytes_per_sec": 0, 00:10:32.461 "w_mbytes_per_sec": 0 00:10:32.461 }, 00:10:32.461 "claimed": true, 00:10:32.461 "claim_type": "exclusive_write", 00:10:32.461 "zoned": false, 00:10:32.461 "supported_io_types": { 00:10:32.461 "read": true, 00:10:32.461 "write": true, 00:10:32.461 "unmap": true, 00:10:32.461 "flush": true, 00:10:32.461 "reset": true, 00:10:32.461 "nvme_admin": false, 00:10:32.461 "nvme_io": false, 00:10:32.461 "nvme_io_md": false, 00:10:32.461 "write_zeroes": true, 00:10:32.461 "zcopy": true, 00:10:32.461 "get_zone_info": false, 00:10:32.461 "zone_management": false, 00:10:32.461 "zone_append": false, 00:10:32.461 "compare": false, 00:10:32.461 "compare_and_write": false, 00:10:32.461 "abort": true, 00:10:32.461 "seek_hole": false, 00:10:32.461 "seek_data": false, 00:10:32.461 "copy": true, 00:10:32.461 "nvme_iov_md": false 00:10:32.461 }, 00:10:32.461 "memory_domains": [ 00:10:32.461 { 00:10:32.461 "dma_device_id": "system", 00:10:32.461 "dma_device_type": 1 00:10:32.461 }, 00:10:32.461 { 00:10:32.461 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.461 "dma_device_type": 2 00:10:32.461 } 00:10:32.461 ], 00:10:32.461 "driver_specific": {} 00:10:32.461 } 00:10:32.461 ] 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.461 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.462 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.462 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:32.462 "name": "Existed_Raid", 00:10:32.462 "uuid": "74f62c4f-f6f0-4cd1-bb8c-803a9baa31f7", 00:10:32.462 "strip_size_kb": 0, 00:10:32.462 "state": "configuring", 00:10:32.462 "raid_level": "raid1", 00:10:32.462 "superblock": true, 00:10:32.462 "num_base_bdevs": 3, 00:10:32.462 "num_base_bdevs_discovered": 1, 00:10:32.462 "num_base_bdevs_operational": 3, 00:10:32.462 "base_bdevs_list": [ 00:10:32.462 { 00:10:32.462 "name": "BaseBdev1", 00:10:32.462 "uuid": "598364e3-d476-441c-99da-b8434c2142d0", 00:10:32.462 "is_configured": true, 00:10:32.462 "data_offset": 2048, 00:10:32.462 "data_size": 63488 00:10:32.462 }, 00:10:32.462 { 00:10:32.462 "name": "BaseBdev2", 00:10:32.462 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.462 "is_configured": false, 00:10:32.462 "data_offset": 0, 00:10:32.462 "data_size": 0 00:10:32.462 }, 00:10:32.462 { 00:10:32.462 "name": "BaseBdev3", 00:10:32.462 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.462 "is_configured": false, 00:10:32.462 "data_offset": 0, 00:10:32.462 "data_size": 0 00:10:32.462 } 00:10:32.462 ] 00:10:32.462 }' 00:10:32.462 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:32.462 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.027 [2024-11-20 15:58:30.984615] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:33.027 [2024-11-20 15:58:30.984657] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.027 [2024-11-20 15:58:30.992656] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:33.027 [2024-11-20 15:58:30.994163] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:33.027 [2024-11-20 15:58:30.994197] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:33.027 [2024-11-20 15:58:30.994205] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:33.027 [2024-11-20 15:58:30.994213] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:33.027 15:58:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.027 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:33.027 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.027 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.027 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.027 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:33.027 "name": "Existed_Raid", 00:10:33.027 "uuid": "e59bb402-650b-4ae9-b399-33be00a3ca8c", 00:10:33.027 "strip_size_kb": 0, 00:10:33.027 "state": "configuring", 00:10:33.027 "raid_level": "raid1", 00:10:33.027 "superblock": true, 00:10:33.027 "num_base_bdevs": 3, 00:10:33.027 "num_base_bdevs_discovered": 1, 00:10:33.027 "num_base_bdevs_operational": 3, 00:10:33.027 "base_bdevs_list": [ 00:10:33.027 { 00:10:33.027 "name": "BaseBdev1", 00:10:33.027 "uuid": "598364e3-d476-441c-99da-b8434c2142d0", 00:10:33.027 "is_configured": true, 00:10:33.027 "data_offset": 2048, 00:10:33.027 "data_size": 63488 00:10:33.027 }, 00:10:33.027 { 00:10:33.027 "name": "BaseBdev2", 00:10:33.027 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:33.027 "is_configured": false, 00:10:33.027 "data_offset": 0, 00:10:33.027 "data_size": 0 00:10:33.027 }, 00:10:33.027 { 00:10:33.027 "name": "BaseBdev3", 00:10:33.027 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:33.027 "is_configured": false, 00:10:33.027 "data_offset": 0, 00:10:33.027 "data_size": 0 00:10:33.027 } 00:10:33.027 ] 00:10:33.027 }' 00:10:33.027 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:33.027 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.285 [2024-11-20 15:58:31.363179] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:33.285 BaseBdev2 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.285 [ 00:10:33.285 { 00:10:33.285 "name": "BaseBdev2", 00:10:33.285 "aliases": [ 00:10:33.285 "f3b1c03d-1f18-4a56-af99-065b9c3ea3c2" 00:10:33.285 ], 00:10:33.285 "product_name": "Malloc disk", 00:10:33.285 "block_size": 512, 00:10:33.285 "num_blocks": 65536, 00:10:33.285 "uuid": "f3b1c03d-1f18-4a56-af99-065b9c3ea3c2", 00:10:33.285 "assigned_rate_limits": { 00:10:33.285 "rw_ios_per_sec": 0, 00:10:33.285 "rw_mbytes_per_sec": 0, 00:10:33.285 "r_mbytes_per_sec": 0, 00:10:33.285 "w_mbytes_per_sec": 0 00:10:33.285 }, 00:10:33.285 "claimed": true, 00:10:33.285 "claim_type": "exclusive_write", 00:10:33.285 "zoned": false, 00:10:33.285 "supported_io_types": { 00:10:33.285 "read": true, 00:10:33.285 "write": true, 00:10:33.285 "unmap": true, 00:10:33.285 "flush": true, 00:10:33.285 "reset": true, 00:10:33.285 "nvme_admin": false, 00:10:33.285 "nvme_io": false, 00:10:33.285 "nvme_io_md": false, 00:10:33.285 "write_zeroes": true, 00:10:33.285 "zcopy": true, 00:10:33.285 "get_zone_info": false, 00:10:33.285 "zone_management": false, 00:10:33.285 "zone_append": false, 00:10:33.285 "compare": false, 00:10:33.285 "compare_and_write": false, 00:10:33.285 "abort": true, 00:10:33.285 "seek_hole": false, 00:10:33.285 "seek_data": false, 00:10:33.285 "copy": true, 00:10:33.285 "nvme_iov_md": false 00:10:33.285 }, 00:10:33.285 "memory_domains": [ 00:10:33.285 { 00:10:33.285 "dma_device_id": "system", 00:10:33.285 "dma_device_type": 1 00:10:33.285 }, 00:10:33.285 { 00:10:33.285 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:33.285 "dma_device_type": 2 00:10:33.285 } 00:10:33.285 ], 00:10:33.285 "driver_specific": {} 00:10:33.285 } 00:10:33.285 ] 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:33.285 "name": "Existed_Raid", 00:10:33.285 "uuid": "e59bb402-650b-4ae9-b399-33be00a3ca8c", 00:10:33.285 "strip_size_kb": 0, 00:10:33.285 "state": "configuring", 00:10:33.285 "raid_level": "raid1", 00:10:33.285 "superblock": true, 00:10:33.285 "num_base_bdevs": 3, 00:10:33.285 "num_base_bdevs_discovered": 2, 00:10:33.285 "num_base_bdevs_operational": 3, 00:10:33.285 "base_bdevs_list": [ 00:10:33.285 { 00:10:33.285 "name": "BaseBdev1", 00:10:33.285 "uuid": "598364e3-d476-441c-99da-b8434c2142d0", 00:10:33.285 "is_configured": true, 00:10:33.285 "data_offset": 2048, 00:10:33.285 "data_size": 63488 00:10:33.285 }, 00:10:33.285 { 00:10:33.285 "name": "BaseBdev2", 00:10:33.285 "uuid": "f3b1c03d-1f18-4a56-af99-065b9c3ea3c2", 00:10:33.285 "is_configured": true, 00:10:33.285 "data_offset": 2048, 00:10:33.285 "data_size": 63488 00:10:33.285 }, 00:10:33.285 { 00:10:33.285 "name": "BaseBdev3", 00:10:33.285 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:33.285 "is_configured": false, 00:10:33.285 "data_offset": 0, 00:10:33.285 "data_size": 0 00:10:33.285 } 00:10:33.285 ] 00:10:33.285 }' 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:33.285 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.543 BaseBdev3 00:10:33.543 [2024-11-20 15:58:31.745554] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:33.543 [2024-11-20 15:58:31.745751] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:10:33.543 [2024-11-20 15:58:31.745767] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:33.543 [2024-11-20 15:58:31.745977] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:10:33.543 [2024-11-20 15:58:31.746083] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:10:33.543 [2024-11-20 15:58:31.746090] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:10:33.543 [2024-11-20 15:58:31.746194] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.543 [ 00:10:33.543 { 00:10:33.543 "name": "BaseBdev3", 00:10:33.543 "aliases": [ 00:10:33.543 "e8e3dce1-128c-4b03-bbd9-3096da9adbda" 00:10:33.543 ], 00:10:33.543 "product_name": "Malloc disk", 00:10:33.543 "block_size": 512, 00:10:33.543 "num_blocks": 65536, 00:10:33.543 "uuid": "e8e3dce1-128c-4b03-bbd9-3096da9adbda", 00:10:33.543 "assigned_rate_limits": { 00:10:33.543 "rw_ios_per_sec": 0, 00:10:33.543 "rw_mbytes_per_sec": 0, 00:10:33.543 "r_mbytes_per_sec": 0, 00:10:33.543 "w_mbytes_per_sec": 0 00:10:33.543 }, 00:10:33.543 "claimed": true, 00:10:33.543 "claim_type": "exclusive_write", 00:10:33.543 "zoned": false, 00:10:33.543 "supported_io_types": { 00:10:33.543 "read": true, 00:10:33.543 "write": true, 00:10:33.543 "unmap": true, 00:10:33.543 "flush": true, 00:10:33.543 "reset": true, 00:10:33.543 "nvme_admin": false, 00:10:33.543 "nvme_io": false, 00:10:33.543 "nvme_io_md": false, 00:10:33.543 "write_zeroes": true, 00:10:33.543 "zcopy": true, 00:10:33.543 "get_zone_info": false, 00:10:33.543 "zone_management": false, 00:10:33.543 "zone_append": false, 00:10:33.543 "compare": false, 00:10:33.543 "compare_and_write": false, 00:10:33.543 "abort": true, 00:10:33.543 "seek_hole": false, 00:10:33.543 "seek_data": false, 00:10:33.543 "copy": true, 00:10:33.543 "nvme_iov_md": false 00:10:33.543 }, 00:10:33.543 "memory_domains": [ 00:10:33.543 { 00:10:33.543 "dma_device_id": "system", 00:10:33.543 "dma_device_type": 1 00:10:33.543 }, 00:10:33.543 { 00:10:33.543 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:33.543 "dma_device_type": 2 00:10:33.543 } 00:10:33.543 ], 00:10:33.543 "driver_specific": {} 00:10:33.543 } 00:10:33.543 ] 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.543 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.801 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:33.801 "name": "Existed_Raid", 00:10:33.801 "uuid": "e59bb402-650b-4ae9-b399-33be00a3ca8c", 00:10:33.801 "strip_size_kb": 0, 00:10:33.801 "state": "online", 00:10:33.801 "raid_level": "raid1", 00:10:33.801 "superblock": true, 00:10:33.801 "num_base_bdevs": 3, 00:10:33.801 "num_base_bdevs_discovered": 3, 00:10:33.801 "num_base_bdevs_operational": 3, 00:10:33.801 "base_bdevs_list": [ 00:10:33.801 { 00:10:33.801 "name": "BaseBdev1", 00:10:33.801 "uuid": "598364e3-d476-441c-99da-b8434c2142d0", 00:10:33.801 "is_configured": true, 00:10:33.801 "data_offset": 2048, 00:10:33.801 "data_size": 63488 00:10:33.801 }, 00:10:33.801 { 00:10:33.801 "name": "BaseBdev2", 00:10:33.801 "uuid": "f3b1c03d-1f18-4a56-af99-065b9c3ea3c2", 00:10:33.801 "is_configured": true, 00:10:33.801 "data_offset": 2048, 00:10:33.801 "data_size": 63488 00:10:33.801 }, 00:10:33.801 { 00:10:33.801 "name": "BaseBdev3", 00:10:33.801 "uuid": "e8e3dce1-128c-4b03-bbd9-3096da9adbda", 00:10:33.801 "is_configured": true, 00:10:33.801 "data_offset": 2048, 00:10:33.801 "data_size": 63488 00:10:33.801 } 00:10:33.801 ] 00:10:33.801 }' 00:10:33.801 15:58:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:33.801 15:58:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.081 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:34.081 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:34.081 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:34.081 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:34.081 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:34.081 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:34.081 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:34.081 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:34.081 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.081 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.081 [2024-11-20 15:58:32.097918] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:34.081 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.081 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:34.081 "name": "Existed_Raid", 00:10:34.081 "aliases": [ 00:10:34.081 "e59bb402-650b-4ae9-b399-33be00a3ca8c" 00:10:34.081 ], 00:10:34.081 "product_name": "Raid Volume", 00:10:34.081 "block_size": 512, 00:10:34.081 "num_blocks": 63488, 00:10:34.081 "uuid": "e59bb402-650b-4ae9-b399-33be00a3ca8c", 00:10:34.081 "assigned_rate_limits": { 00:10:34.081 "rw_ios_per_sec": 0, 00:10:34.081 "rw_mbytes_per_sec": 0, 00:10:34.081 "r_mbytes_per_sec": 0, 00:10:34.081 "w_mbytes_per_sec": 0 00:10:34.081 }, 00:10:34.081 "claimed": false, 00:10:34.081 "zoned": false, 00:10:34.081 "supported_io_types": { 00:10:34.081 "read": true, 00:10:34.081 "write": true, 00:10:34.081 "unmap": false, 00:10:34.081 "flush": false, 00:10:34.081 "reset": true, 00:10:34.081 "nvme_admin": false, 00:10:34.081 "nvme_io": false, 00:10:34.081 "nvme_io_md": false, 00:10:34.081 "write_zeroes": true, 00:10:34.081 "zcopy": false, 00:10:34.081 "get_zone_info": false, 00:10:34.081 "zone_management": false, 00:10:34.081 "zone_append": false, 00:10:34.081 "compare": false, 00:10:34.081 "compare_and_write": false, 00:10:34.081 "abort": false, 00:10:34.081 "seek_hole": false, 00:10:34.081 "seek_data": false, 00:10:34.081 "copy": false, 00:10:34.081 "nvme_iov_md": false 00:10:34.081 }, 00:10:34.081 "memory_domains": [ 00:10:34.081 { 00:10:34.082 "dma_device_id": "system", 00:10:34.082 "dma_device_type": 1 00:10:34.082 }, 00:10:34.082 { 00:10:34.082 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.082 "dma_device_type": 2 00:10:34.082 }, 00:10:34.082 { 00:10:34.082 "dma_device_id": "system", 00:10:34.082 "dma_device_type": 1 00:10:34.082 }, 00:10:34.082 { 00:10:34.082 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.082 "dma_device_type": 2 00:10:34.082 }, 00:10:34.082 { 00:10:34.082 "dma_device_id": "system", 00:10:34.082 "dma_device_type": 1 00:10:34.082 }, 00:10:34.082 { 00:10:34.082 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.082 "dma_device_type": 2 00:10:34.082 } 00:10:34.082 ], 00:10:34.082 "driver_specific": { 00:10:34.082 "raid": { 00:10:34.082 "uuid": "e59bb402-650b-4ae9-b399-33be00a3ca8c", 00:10:34.082 "strip_size_kb": 0, 00:10:34.082 "state": "online", 00:10:34.082 "raid_level": "raid1", 00:10:34.082 "superblock": true, 00:10:34.082 "num_base_bdevs": 3, 00:10:34.082 "num_base_bdevs_discovered": 3, 00:10:34.082 "num_base_bdevs_operational": 3, 00:10:34.082 "base_bdevs_list": [ 00:10:34.082 { 00:10:34.082 "name": "BaseBdev1", 00:10:34.082 "uuid": "598364e3-d476-441c-99da-b8434c2142d0", 00:10:34.082 "is_configured": true, 00:10:34.082 "data_offset": 2048, 00:10:34.082 "data_size": 63488 00:10:34.082 }, 00:10:34.082 { 00:10:34.082 "name": "BaseBdev2", 00:10:34.082 "uuid": "f3b1c03d-1f18-4a56-af99-065b9c3ea3c2", 00:10:34.082 "is_configured": true, 00:10:34.082 "data_offset": 2048, 00:10:34.082 "data_size": 63488 00:10:34.082 }, 00:10:34.082 { 00:10:34.082 "name": "BaseBdev3", 00:10:34.082 "uuid": "e8e3dce1-128c-4b03-bbd9-3096da9adbda", 00:10:34.082 "is_configured": true, 00:10:34.082 "data_offset": 2048, 00:10:34.082 "data_size": 63488 00:10:34.082 } 00:10:34.082 ] 00:10:34.082 } 00:10:34.082 } 00:10:34.082 }' 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:34.082 BaseBdev2 00:10:34.082 BaseBdev3' 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.082 [2024-11-20 15:58:32.277745] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:34.082 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:34.339 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.339 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:34.339 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.339 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.339 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.340 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:34.340 "name": "Existed_Raid", 00:10:34.340 "uuid": "e59bb402-650b-4ae9-b399-33be00a3ca8c", 00:10:34.340 "strip_size_kb": 0, 00:10:34.340 "state": "online", 00:10:34.340 "raid_level": "raid1", 00:10:34.340 "superblock": true, 00:10:34.340 "num_base_bdevs": 3, 00:10:34.340 "num_base_bdevs_discovered": 2, 00:10:34.340 "num_base_bdevs_operational": 2, 00:10:34.340 "base_bdevs_list": [ 00:10:34.340 { 00:10:34.340 "name": null, 00:10:34.340 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:34.340 "is_configured": false, 00:10:34.340 "data_offset": 0, 00:10:34.340 "data_size": 63488 00:10:34.340 }, 00:10:34.340 { 00:10:34.340 "name": "BaseBdev2", 00:10:34.340 "uuid": "f3b1c03d-1f18-4a56-af99-065b9c3ea3c2", 00:10:34.340 "is_configured": true, 00:10:34.340 "data_offset": 2048, 00:10:34.340 "data_size": 63488 00:10:34.340 }, 00:10:34.340 { 00:10:34.340 "name": "BaseBdev3", 00:10:34.340 "uuid": "e8e3dce1-128c-4b03-bbd9-3096da9adbda", 00:10:34.340 "is_configured": true, 00:10:34.340 "data_offset": 2048, 00:10:34.340 "data_size": 63488 00:10:34.340 } 00:10:34.340 ] 00:10:34.340 }' 00:10:34.340 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:34.340 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.597 [2024-11-20 15:58:32.667419] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.597 [2024-11-20 15:58:32.745946] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:34.597 [2024-11-20 15:58:32.746024] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:34.597 [2024-11-20 15:58:32.792577] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:34.597 [2024-11-20 15:58:32.792739] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:34.597 [2024-11-20 15:58:32.792802] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.597 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.855 BaseBdev2 00:10:34.855 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.855 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:34.855 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:34.855 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:34.855 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:34.855 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:34.855 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:34.855 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:34.855 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.855 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.855 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.856 [ 00:10:34.856 { 00:10:34.856 "name": "BaseBdev2", 00:10:34.856 "aliases": [ 00:10:34.856 "a10b0973-6269-4db4-9328-9c651a197a3c" 00:10:34.856 ], 00:10:34.856 "product_name": "Malloc disk", 00:10:34.856 "block_size": 512, 00:10:34.856 "num_blocks": 65536, 00:10:34.856 "uuid": "a10b0973-6269-4db4-9328-9c651a197a3c", 00:10:34.856 "assigned_rate_limits": { 00:10:34.856 "rw_ios_per_sec": 0, 00:10:34.856 "rw_mbytes_per_sec": 0, 00:10:34.856 "r_mbytes_per_sec": 0, 00:10:34.856 "w_mbytes_per_sec": 0 00:10:34.856 }, 00:10:34.856 "claimed": false, 00:10:34.856 "zoned": false, 00:10:34.856 "supported_io_types": { 00:10:34.856 "read": true, 00:10:34.856 "write": true, 00:10:34.856 "unmap": true, 00:10:34.856 "flush": true, 00:10:34.856 "reset": true, 00:10:34.856 "nvme_admin": false, 00:10:34.856 "nvme_io": false, 00:10:34.856 "nvme_io_md": false, 00:10:34.856 "write_zeroes": true, 00:10:34.856 "zcopy": true, 00:10:34.856 "get_zone_info": false, 00:10:34.856 "zone_management": false, 00:10:34.856 "zone_append": false, 00:10:34.856 "compare": false, 00:10:34.856 "compare_and_write": false, 00:10:34.856 "abort": true, 00:10:34.856 "seek_hole": false, 00:10:34.856 "seek_data": false, 00:10:34.856 "copy": true, 00:10:34.856 "nvme_iov_md": false 00:10:34.856 }, 00:10:34.856 "memory_domains": [ 00:10:34.856 { 00:10:34.856 "dma_device_id": "system", 00:10:34.856 "dma_device_type": 1 00:10:34.856 }, 00:10:34.856 { 00:10:34.856 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.856 "dma_device_type": 2 00:10:34.856 } 00:10:34.856 ], 00:10:34.856 "driver_specific": {} 00:10:34.856 } 00:10:34.856 ] 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.856 BaseBdev3 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.856 [ 00:10:34.856 { 00:10:34.856 "name": "BaseBdev3", 00:10:34.856 "aliases": [ 00:10:34.856 "adc5bade-6652-4aaf-8905-e1b5820a0bdc" 00:10:34.856 ], 00:10:34.856 "product_name": "Malloc disk", 00:10:34.856 "block_size": 512, 00:10:34.856 "num_blocks": 65536, 00:10:34.856 "uuid": "adc5bade-6652-4aaf-8905-e1b5820a0bdc", 00:10:34.856 "assigned_rate_limits": { 00:10:34.856 "rw_ios_per_sec": 0, 00:10:34.856 "rw_mbytes_per_sec": 0, 00:10:34.856 "r_mbytes_per_sec": 0, 00:10:34.856 "w_mbytes_per_sec": 0 00:10:34.856 }, 00:10:34.856 "claimed": false, 00:10:34.856 "zoned": false, 00:10:34.856 "supported_io_types": { 00:10:34.856 "read": true, 00:10:34.856 "write": true, 00:10:34.856 "unmap": true, 00:10:34.856 "flush": true, 00:10:34.856 "reset": true, 00:10:34.856 "nvme_admin": false, 00:10:34.856 "nvme_io": false, 00:10:34.856 "nvme_io_md": false, 00:10:34.856 "write_zeroes": true, 00:10:34.856 "zcopy": true, 00:10:34.856 "get_zone_info": false, 00:10:34.856 "zone_management": false, 00:10:34.856 "zone_append": false, 00:10:34.856 "compare": false, 00:10:34.856 "compare_and_write": false, 00:10:34.856 "abort": true, 00:10:34.856 "seek_hole": false, 00:10:34.856 "seek_data": false, 00:10:34.856 "copy": true, 00:10:34.856 "nvme_iov_md": false 00:10:34.856 }, 00:10:34.856 "memory_domains": [ 00:10:34.856 { 00:10:34.856 "dma_device_id": "system", 00:10:34.856 "dma_device_type": 1 00:10:34.856 }, 00:10:34.856 { 00:10:34.856 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.856 "dma_device_type": 2 00:10:34.856 } 00:10:34.856 ], 00:10:34.856 "driver_specific": {} 00:10:34.856 } 00:10:34.856 ] 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.856 [2024-11-20 15:58:32.940313] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:34.856 [2024-11-20 15:58:32.940434] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:34.856 [2024-11-20 15:58:32.940490] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:34.856 [2024-11-20 15:58:32.941974] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:34.856 "name": "Existed_Raid", 00:10:34.856 "uuid": "f8f8487e-b9c7-45a9-970e-a6bfa1b4821f", 00:10:34.856 "strip_size_kb": 0, 00:10:34.856 "state": "configuring", 00:10:34.856 "raid_level": "raid1", 00:10:34.856 "superblock": true, 00:10:34.856 "num_base_bdevs": 3, 00:10:34.856 "num_base_bdevs_discovered": 2, 00:10:34.856 "num_base_bdevs_operational": 3, 00:10:34.856 "base_bdevs_list": [ 00:10:34.856 { 00:10:34.856 "name": "BaseBdev1", 00:10:34.856 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:34.856 "is_configured": false, 00:10:34.856 "data_offset": 0, 00:10:34.856 "data_size": 0 00:10:34.856 }, 00:10:34.856 { 00:10:34.856 "name": "BaseBdev2", 00:10:34.856 "uuid": "a10b0973-6269-4db4-9328-9c651a197a3c", 00:10:34.856 "is_configured": true, 00:10:34.856 "data_offset": 2048, 00:10:34.856 "data_size": 63488 00:10:34.856 }, 00:10:34.856 { 00:10:34.856 "name": "BaseBdev3", 00:10:34.856 "uuid": "adc5bade-6652-4aaf-8905-e1b5820a0bdc", 00:10:34.856 "is_configured": true, 00:10:34.856 "data_offset": 2048, 00:10:34.856 "data_size": 63488 00:10:34.856 } 00:10:34.856 ] 00:10:34.856 }' 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:34.856 15:58:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.122 [2024-11-20 15:58:33.260393] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.122 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:35.122 "name": "Existed_Raid", 00:10:35.122 "uuid": "f8f8487e-b9c7-45a9-970e-a6bfa1b4821f", 00:10:35.122 "strip_size_kb": 0, 00:10:35.122 "state": "configuring", 00:10:35.122 "raid_level": "raid1", 00:10:35.122 "superblock": true, 00:10:35.122 "num_base_bdevs": 3, 00:10:35.122 "num_base_bdevs_discovered": 1, 00:10:35.122 "num_base_bdevs_operational": 3, 00:10:35.122 "base_bdevs_list": [ 00:10:35.122 { 00:10:35.122 "name": "BaseBdev1", 00:10:35.123 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:35.123 "is_configured": false, 00:10:35.123 "data_offset": 0, 00:10:35.123 "data_size": 0 00:10:35.123 }, 00:10:35.123 { 00:10:35.123 "name": null, 00:10:35.123 "uuid": "a10b0973-6269-4db4-9328-9c651a197a3c", 00:10:35.123 "is_configured": false, 00:10:35.123 "data_offset": 0, 00:10:35.123 "data_size": 63488 00:10:35.123 }, 00:10:35.123 { 00:10:35.123 "name": "BaseBdev3", 00:10:35.123 "uuid": "adc5bade-6652-4aaf-8905-e1b5820a0bdc", 00:10:35.123 "is_configured": true, 00:10:35.123 "data_offset": 2048, 00:10:35.123 "data_size": 63488 00:10:35.123 } 00:10:35.123 ] 00:10:35.123 }' 00:10:35.123 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:35.123 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.379 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.379 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:35.379 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.379 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.379 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.379 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:35.379 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:35.379 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.379 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.635 [2024-11-20 15:58:33.646967] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:35.635 BaseBdev1 00:10:35.635 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.635 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:35.635 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:35.635 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:35.635 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:35.635 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:35.635 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:35.635 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:35.635 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.635 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.635 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.635 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:35.635 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.635 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.635 [ 00:10:35.635 { 00:10:35.635 "name": "BaseBdev1", 00:10:35.635 "aliases": [ 00:10:35.635 "b6822a16-cac8-45db-b725-c9a4f175847a" 00:10:35.635 ], 00:10:35.635 "product_name": "Malloc disk", 00:10:35.635 "block_size": 512, 00:10:35.635 "num_blocks": 65536, 00:10:35.635 "uuid": "b6822a16-cac8-45db-b725-c9a4f175847a", 00:10:35.635 "assigned_rate_limits": { 00:10:35.635 "rw_ios_per_sec": 0, 00:10:35.635 "rw_mbytes_per_sec": 0, 00:10:35.635 "r_mbytes_per_sec": 0, 00:10:35.635 "w_mbytes_per_sec": 0 00:10:35.635 }, 00:10:35.635 "claimed": true, 00:10:35.635 "claim_type": "exclusive_write", 00:10:35.635 "zoned": false, 00:10:35.635 "supported_io_types": { 00:10:35.635 "read": true, 00:10:35.635 "write": true, 00:10:35.635 "unmap": true, 00:10:35.635 "flush": true, 00:10:35.635 "reset": true, 00:10:35.635 "nvme_admin": false, 00:10:35.635 "nvme_io": false, 00:10:35.635 "nvme_io_md": false, 00:10:35.635 "write_zeroes": true, 00:10:35.635 "zcopy": true, 00:10:35.635 "get_zone_info": false, 00:10:35.635 "zone_management": false, 00:10:35.635 "zone_append": false, 00:10:35.635 "compare": false, 00:10:35.635 "compare_and_write": false, 00:10:35.635 "abort": true, 00:10:35.635 "seek_hole": false, 00:10:35.635 "seek_data": false, 00:10:35.635 "copy": true, 00:10:35.635 "nvme_iov_md": false 00:10:35.635 }, 00:10:35.635 "memory_domains": [ 00:10:35.635 { 00:10:35.635 "dma_device_id": "system", 00:10:35.636 "dma_device_type": 1 00:10:35.636 }, 00:10:35.636 { 00:10:35.636 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:35.636 "dma_device_type": 2 00:10:35.636 } 00:10:35.636 ], 00:10:35.636 "driver_specific": {} 00:10:35.636 } 00:10:35.636 ] 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:35.636 "name": "Existed_Raid", 00:10:35.636 "uuid": "f8f8487e-b9c7-45a9-970e-a6bfa1b4821f", 00:10:35.636 "strip_size_kb": 0, 00:10:35.636 "state": "configuring", 00:10:35.636 "raid_level": "raid1", 00:10:35.636 "superblock": true, 00:10:35.636 "num_base_bdevs": 3, 00:10:35.636 "num_base_bdevs_discovered": 2, 00:10:35.636 "num_base_bdevs_operational": 3, 00:10:35.636 "base_bdevs_list": [ 00:10:35.636 { 00:10:35.636 "name": "BaseBdev1", 00:10:35.636 "uuid": "b6822a16-cac8-45db-b725-c9a4f175847a", 00:10:35.636 "is_configured": true, 00:10:35.636 "data_offset": 2048, 00:10:35.636 "data_size": 63488 00:10:35.636 }, 00:10:35.636 { 00:10:35.636 "name": null, 00:10:35.636 "uuid": "a10b0973-6269-4db4-9328-9c651a197a3c", 00:10:35.636 "is_configured": false, 00:10:35.636 "data_offset": 0, 00:10:35.636 "data_size": 63488 00:10:35.636 }, 00:10:35.636 { 00:10:35.636 "name": "BaseBdev3", 00:10:35.636 "uuid": "adc5bade-6652-4aaf-8905-e1b5820a0bdc", 00:10:35.636 "is_configured": true, 00:10:35.636 "data_offset": 2048, 00:10:35.636 "data_size": 63488 00:10:35.636 } 00:10:35.636 ] 00:10:35.636 }' 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:35.636 15:58:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.892 [2024-11-20 15:58:34.043070] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:35.892 "name": "Existed_Raid", 00:10:35.892 "uuid": "f8f8487e-b9c7-45a9-970e-a6bfa1b4821f", 00:10:35.892 "strip_size_kb": 0, 00:10:35.892 "state": "configuring", 00:10:35.892 "raid_level": "raid1", 00:10:35.892 "superblock": true, 00:10:35.892 "num_base_bdevs": 3, 00:10:35.892 "num_base_bdevs_discovered": 1, 00:10:35.892 "num_base_bdevs_operational": 3, 00:10:35.892 "base_bdevs_list": [ 00:10:35.892 { 00:10:35.892 "name": "BaseBdev1", 00:10:35.892 "uuid": "b6822a16-cac8-45db-b725-c9a4f175847a", 00:10:35.892 "is_configured": true, 00:10:35.892 "data_offset": 2048, 00:10:35.892 "data_size": 63488 00:10:35.892 }, 00:10:35.892 { 00:10:35.892 "name": null, 00:10:35.892 "uuid": "a10b0973-6269-4db4-9328-9c651a197a3c", 00:10:35.892 "is_configured": false, 00:10:35.892 "data_offset": 0, 00:10:35.892 "data_size": 63488 00:10:35.892 }, 00:10:35.892 { 00:10:35.892 "name": null, 00:10:35.892 "uuid": "adc5bade-6652-4aaf-8905-e1b5820a0bdc", 00:10:35.892 "is_configured": false, 00:10:35.892 "data_offset": 0, 00:10:35.892 "data_size": 63488 00:10:35.892 } 00:10:35.892 ] 00:10:35.892 }' 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:35.892 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.150 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.150 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.150 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.150 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:36.150 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.150 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:36.150 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:36.150 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.150 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.407 [2024-11-20 15:58:34.399178] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:36.407 "name": "Existed_Raid", 00:10:36.407 "uuid": "f8f8487e-b9c7-45a9-970e-a6bfa1b4821f", 00:10:36.407 "strip_size_kb": 0, 00:10:36.407 "state": "configuring", 00:10:36.407 "raid_level": "raid1", 00:10:36.407 "superblock": true, 00:10:36.407 "num_base_bdevs": 3, 00:10:36.407 "num_base_bdevs_discovered": 2, 00:10:36.407 "num_base_bdevs_operational": 3, 00:10:36.407 "base_bdevs_list": [ 00:10:36.407 { 00:10:36.407 "name": "BaseBdev1", 00:10:36.407 "uuid": "b6822a16-cac8-45db-b725-c9a4f175847a", 00:10:36.407 "is_configured": true, 00:10:36.407 "data_offset": 2048, 00:10:36.407 "data_size": 63488 00:10:36.407 }, 00:10:36.407 { 00:10:36.407 "name": null, 00:10:36.407 "uuid": "a10b0973-6269-4db4-9328-9c651a197a3c", 00:10:36.407 "is_configured": false, 00:10:36.407 "data_offset": 0, 00:10:36.407 "data_size": 63488 00:10:36.407 }, 00:10:36.407 { 00:10:36.407 "name": "BaseBdev3", 00:10:36.407 "uuid": "adc5bade-6652-4aaf-8905-e1b5820a0bdc", 00:10:36.407 "is_configured": true, 00:10:36.407 "data_offset": 2048, 00:10:36.407 "data_size": 63488 00:10:36.407 } 00:10:36.407 ] 00:10:36.407 }' 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:36.407 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.664 [2024-11-20 15:58:34.747256] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:36.664 "name": "Existed_Raid", 00:10:36.664 "uuid": "f8f8487e-b9c7-45a9-970e-a6bfa1b4821f", 00:10:36.664 "strip_size_kb": 0, 00:10:36.664 "state": "configuring", 00:10:36.664 "raid_level": "raid1", 00:10:36.664 "superblock": true, 00:10:36.664 "num_base_bdevs": 3, 00:10:36.664 "num_base_bdevs_discovered": 1, 00:10:36.664 "num_base_bdevs_operational": 3, 00:10:36.664 "base_bdevs_list": [ 00:10:36.664 { 00:10:36.664 "name": null, 00:10:36.664 "uuid": "b6822a16-cac8-45db-b725-c9a4f175847a", 00:10:36.664 "is_configured": false, 00:10:36.664 "data_offset": 0, 00:10:36.664 "data_size": 63488 00:10:36.664 }, 00:10:36.664 { 00:10:36.664 "name": null, 00:10:36.664 "uuid": "a10b0973-6269-4db4-9328-9c651a197a3c", 00:10:36.664 "is_configured": false, 00:10:36.664 "data_offset": 0, 00:10:36.664 "data_size": 63488 00:10:36.664 }, 00:10:36.664 { 00:10:36.664 "name": "BaseBdev3", 00:10:36.664 "uuid": "adc5bade-6652-4aaf-8905-e1b5820a0bdc", 00:10:36.664 "is_configured": true, 00:10:36.664 "data_offset": 2048, 00:10:36.664 "data_size": 63488 00:10:36.664 } 00:10:36.664 ] 00:10:36.664 }' 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:36.664 15:58:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.922 [2024-11-20 15:58:35.131327] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.922 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.179 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:37.179 "name": "Existed_Raid", 00:10:37.179 "uuid": "f8f8487e-b9c7-45a9-970e-a6bfa1b4821f", 00:10:37.179 "strip_size_kb": 0, 00:10:37.179 "state": "configuring", 00:10:37.179 "raid_level": "raid1", 00:10:37.179 "superblock": true, 00:10:37.179 "num_base_bdevs": 3, 00:10:37.179 "num_base_bdevs_discovered": 2, 00:10:37.179 "num_base_bdevs_operational": 3, 00:10:37.179 "base_bdevs_list": [ 00:10:37.179 { 00:10:37.179 "name": null, 00:10:37.179 "uuid": "b6822a16-cac8-45db-b725-c9a4f175847a", 00:10:37.179 "is_configured": false, 00:10:37.179 "data_offset": 0, 00:10:37.179 "data_size": 63488 00:10:37.179 }, 00:10:37.179 { 00:10:37.179 "name": "BaseBdev2", 00:10:37.179 "uuid": "a10b0973-6269-4db4-9328-9c651a197a3c", 00:10:37.179 "is_configured": true, 00:10:37.179 "data_offset": 2048, 00:10:37.179 "data_size": 63488 00:10:37.179 }, 00:10:37.179 { 00:10:37.179 "name": "BaseBdev3", 00:10:37.179 "uuid": "adc5bade-6652-4aaf-8905-e1b5820a0bdc", 00:10:37.179 "is_configured": true, 00:10:37.179 "data_offset": 2048, 00:10:37.179 "data_size": 63488 00:10:37.179 } 00:10:37.179 ] 00:10:37.179 }' 00:10:37.179 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:37.179 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.436 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.436 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.436 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:37.436 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.436 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.436 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u b6822a16-cac8-45db-b725-c9a4f175847a 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.437 [2024-11-20 15:58:35.554182] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:37.437 [2024-11-20 15:58:35.554374] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:10:37.437 [2024-11-20 15:58:35.554384] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:37.437 [2024-11-20 15:58:35.554582] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:10:37.437 NewBaseBdev 00:10:37.437 [2024-11-20 15:58:35.554698] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:10:37.437 [2024-11-20 15:58:35.554708] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:10:37.437 [2024-11-20 15:58:35.554802] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.437 [ 00:10:37.437 { 00:10:37.437 "name": "NewBaseBdev", 00:10:37.437 "aliases": [ 00:10:37.437 "b6822a16-cac8-45db-b725-c9a4f175847a" 00:10:37.437 ], 00:10:37.437 "product_name": "Malloc disk", 00:10:37.437 "block_size": 512, 00:10:37.437 "num_blocks": 65536, 00:10:37.437 "uuid": "b6822a16-cac8-45db-b725-c9a4f175847a", 00:10:37.437 "assigned_rate_limits": { 00:10:37.437 "rw_ios_per_sec": 0, 00:10:37.437 "rw_mbytes_per_sec": 0, 00:10:37.437 "r_mbytes_per_sec": 0, 00:10:37.437 "w_mbytes_per_sec": 0 00:10:37.437 }, 00:10:37.437 "claimed": true, 00:10:37.437 "claim_type": "exclusive_write", 00:10:37.437 "zoned": false, 00:10:37.437 "supported_io_types": { 00:10:37.437 "read": true, 00:10:37.437 "write": true, 00:10:37.437 "unmap": true, 00:10:37.437 "flush": true, 00:10:37.437 "reset": true, 00:10:37.437 "nvme_admin": false, 00:10:37.437 "nvme_io": false, 00:10:37.437 "nvme_io_md": false, 00:10:37.437 "write_zeroes": true, 00:10:37.437 "zcopy": true, 00:10:37.437 "get_zone_info": false, 00:10:37.437 "zone_management": false, 00:10:37.437 "zone_append": false, 00:10:37.437 "compare": false, 00:10:37.437 "compare_and_write": false, 00:10:37.437 "abort": true, 00:10:37.437 "seek_hole": false, 00:10:37.437 "seek_data": false, 00:10:37.437 "copy": true, 00:10:37.437 "nvme_iov_md": false 00:10:37.437 }, 00:10:37.437 "memory_domains": [ 00:10:37.437 { 00:10:37.437 "dma_device_id": "system", 00:10:37.437 "dma_device_type": 1 00:10:37.437 }, 00:10:37.437 { 00:10:37.437 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.437 "dma_device_type": 2 00:10:37.437 } 00:10:37.437 ], 00:10:37.437 "driver_specific": {} 00:10:37.437 } 00:10:37.437 ] 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:37.437 "name": "Existed_Raid", 00:10:37.437 "uuid": "f8f8487e-b9c7-45a9-970e-a6bfa1b4821f", 00:10:37.437 "strip_size_kb": 0, 00:10:37.437 "state": "online", 00:10:37.437 "raid_level": "raid1", 00:10:37.437 "superblock": true, 00:10:37.437 "num_base_bdevs": 3, 00:10:37.437 "num_base_bdevs_discovered": 3, 00:10:37.437 "num_base_bdevs_operational": 3, 00:10:37.437 "base_bdevs_list": [ 00:10:37.437 { 00:10:37.437 "name": "NewBaseBdev", 00:10:37.437 "uuid": "b6822a16-cac8-45db-b725-c9a4f175847a", 00:10:37.437 "is_configured": true, 00:10:37.437 "data_offset": 2048, 00:10:37.437 "data_size": 63488 00:10:37.437 }, 00:10:37.437 { 00:10:37.437 "name": "BaseBdev2", 00:10:37.437 "uuid": "a10b0973-6269-4db4-9328-9c651a197a3c", 00:10:37.437 "is_configured": true, 00:10:37.437 "data_offset": 2048, 00:10:37.437 "data_size": 63488 00:10:37.437 }, 00:10:37.437 { 00:10:37.437 "name": "BaseBdev3", 00:10:37.437 "uuid": "adc5bade-6652-4aaf-8905-e1b5820a0bdc", 00:10:37.437 "is_configured": true, 00:10:37.437 "data_offset": 2048, 00:10:37.437 "data_size": 63488 00:10:37.437 } 00:10:37.437 ] 00:10:37.437 }' 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:37.437 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.694 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:37.694 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:37.694 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:37.694 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:37.694 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:37.694 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:37.694 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:37.694 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:37.694 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.694 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.694 [2024-11-20 15:58:35.890544] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:37.694 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.694 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:37.694 "name": "Existed_Raid", 00:10:37.694 "aliases": [ 00:10:37.694 "f8f8487e-b9c7-45a9-970e-a6bfa1b4821f" 00:10:37.694 ], 00:10:37.694 "product_name": "Raid Volume", 00:10:37.694 "block_size": 512, 00:10:37.694 "num_blocks": 63488, 00:10:37.694 "uuid": "f8f8487e-b9c7-45a9-970e-a6bfa1b4821f", 00:10:37.694 "assigned_rate_limits": { 00:10:37.694 "rw_ios_per_sec": 0, 00:10:37.694 "rw_mbytes_per_sec": 0, 00:10:37.694 "r_mbytes_per_sec": 0, 00:10:37.694 "w_mbytes_per_sec": 0 00:10:37.694 }, 00:10:37.694 "claimed": false, 00:10:37.694 "zoned": false, 00:10:37.694 "supported_io_types": { 00:10:37.694 "read": true, 00:10:37.694 "write": true, 00:10:37.694 "unmap": false, 00:10:37.694 "flush": false, 00:10:37.694 "reset": true, 00:10:37.694 "nvme_admin": false, 00:10:37.694 "nvme_io": false, 00:10:37.694 "nvme_io_md": false, 00:10:37.694 "write_zeroes": true, 00:10:37.694 "zcopy": false, 00:10:37.694 "get_zone_info": false, 00:10:37.694 "zone_management": false, 00:10:37.694 "zone_append": false, 00:10:37.694 "compare": false, 00:10:37.694 "compare_and_write": false, 00:10:37.694 "abort": false, 00:10:37.694 "seek_hole": false, 00:10:37.694 "seek_data": false, 00:10:37.694 "copy": false, 00:10:37.694 "nvme_iov_md": false 00:10:37.694 }, 00:10:37.694 "memory_domains": [ 00:10:37.694 { 00:10:37.694 "dma_device_id": "system", 00:10:37.694 "dma_device_type": 1 00:10:37.694 }, 00:10:37.694 { 00:10:37.694 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.694 "dma_device_type": 2 00:10:37.694 }, 00:10:37.694 { 00:10:37.694 "dma_device_id": "system", 00:10:37.694 "dma_device_type": 1 00:10:37.694 }, 00:10:37.694 { 00:10:37.694 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.694 "dma_device_type": 2 00:10:37.694 }, 00:10:37.694 { 00:10:37.694 "dma_device_id": "system", 00:10:37.694 "dma_device_type": 1 00:10:37.694 }, 00:10:37.694 { 00:10:37.694 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.694 "dma_device_type": 2 00:10:37.694 } 00:10:37.694 ], 00:10:37.694 "driver_specific": { 00:10:37.694 "raid": { 00:10:37.694 "uuid": "f8f8487e-b9c7-45a9-970e-a6bfa1b4821f", 00:10:37.694 "strip_size_kb": 0, 00:10:37.694 "state": "online", 00:10:37.694 "raid_level": "raid1", 00:10:37.694 "superblock": true, 00:10:37.694 "num_base_bdevs": 3, 00:10:37.694 "num_base_bdevs_discovered": 3, 00:10:37.694 "num_base_bdevs_operational": 3, 00:10:37.694 "base_bdevs_list": [ 00:10:37.694 { 00:10:37.694 "name": "NewBaseBdev", 00:10:37.694 "uuid": "b6822a16-cac8-45db-b725-c9a4f175847a", 00:10:37.694 "is_configured": true, 00:10:37.694 "data_offset": 2048, 00:10:37.694 "data_size": 63488 00:10:37.694 }, 00:10:37.694 { 00:10:37.694 "name": "BaseBdev2", 00:10:37.694 "uuid": "a10b0973-6269-4db4-9328-9c651a197a3c", 00:10:37.694 "is_configured": true, 00:10:37.694 "data_offset": 2048, 00:10:37.694 "data_size": 63488 00:10:37.694 }, 00:10:37.694 { 00:10:37.694 "name": "BaseBdev3", 00:10:37.694 "uuid": "adc5bade-6652-4aaf-8905-e1b5820a0bdc", 00:10:37.694 "is_configured": true, 00:10:37.694 "data_offset": 2048, 00:10:37.694 "data_size": 63488 00:10:37.694 } 00:10:37.694 ] 00:10:37.694 } 00:10:37.694 } 00:10:37.694 }' 00:10:37.694 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:37.951 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:37.951 BaseBdev2 00:10:37.951 BaseBdev3' 00:10:37.951 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.951 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:37.951 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:37.951 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:37.951 15:58:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.951 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.951 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.951 15:58:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.951 15:58:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:37.951 15:58:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:37.951 15:58:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:37.951 15:58:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:37.951 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.951 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.951 15:58:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.951 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.952 [2024-11-20 15:58:36.082319] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:37.952 [2024-11-20 15:58:36.082345] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:37.952 [2024-11-20 15:58:36.082402] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:37.952 [2024-11-20 15:58:36.082625] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:37.952 [2024-11-20 15:58:36.082633] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 66290 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 66290 ']' 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 66290 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 66290 00:10:37.952 killing process with pid 66290 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 66290' 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 66290 00:10:37.952 [2024-11-20 15:58:36.114692] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:37.952 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 66290 00:10:38.209 [2024-11-20 15:58:36.263000] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:38.772 15:58:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:38.772 00:10:38.772 real 0m7.467s 00:10:38.772 user 0m12.145s 00:10:38.772 sys 0m1.170s 00:10:38.772 ************************************ 00:10:38.772 END TEST raid_state_function_test_sb 00:10:38.772 ************************************ 00:10:38.773 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:38.773 15:58:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:38.773 15:58:36 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 3 00:10:38.773 15:58:36 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:10:38.773 15:58:36 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:38.773 15:58:36 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:38.773 ************************************ 00:10:38.773 START TEST raid_superblock_test 00:10:38.773 ************************************ 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 3 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:38.773 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=66883 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 66883 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 66883 ']' 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:38.773 15:58:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.773 [2024-11-20 15:58:36.935282] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:10:38.773 [2024-11-20 15:58:36.935572] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid66883 ] 00:10:39.030 [2024-11-20 15:58:37.095214] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:39.030 [2024-11-20 15:58:37.197696] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:39.286 [2024-11-20 15:58:37.342853] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:39.286 [2024-11-20 15:58:37.343082] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.542 malloc1 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.542 [2024-11-20 15:58:37.763268] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:39.542 [2024-11-20 15:58:37.763314] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:39.542 [2024-11-20 15:58:37.763332] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:39.542 [2024-11-20 15:58:37.763340] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:39.542 [2024-11-20 15:58:37.765120] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:39.542 [2024-11-20 15:58:37.765146] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:39.542 pt1 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.542 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.800 malloc2 00:10:39.800 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.800 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:39.800 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.800 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.800 [2024-11-20 15:58:37.794988] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:39.800 [2024-11-20 15:58:37.795035] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:39.800 [2024-11-20 15:58:37.795054] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:39.800 [2024-11-20 15:58:37.795060] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:39.801 [2024-11-20 15:58:37.796812] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:39.801 [2024-11-20 15:58:37.796937] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:39.801 pt2 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.801 malloc3 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.801 [2024-11-20 15:58:37.841585] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:39.801 [2024-11-20 15:58:37.841637] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:39.801 [2024-11-20 15:58:37.841654] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:39.801 [2024-11-20 15:58:37.841662] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:39.801 [2024-11-20 15:58:37.843420] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:39.801 [2024-11-20 15:58:37.843561] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:39.801 pt3 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.801 [2024-11-20 15:58:37.849622] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:39.801 [2024-11-20 15:58:37.851234] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:39.801 [2024-11-20 15:58:37.851356] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:39.801 [2024-11-20 15:58:37.851540] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:10:39.801 [2024-11-20 15:58:37.851606] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:39.801 [2024-11-20 15:58:37.851843] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:10:39.801 [2024-11-20 15:58:37.852027] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:10:39.801 [2024-11-20 15:58:37.852083] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:10:39.801 [2024-11-20 15:58:37.852272] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:39.801 "name": "raid_bdev1", 00:10:39.801 "uuid": "eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f", 00:10:39.801 "strip_size_kb": 0, 00:10:39.801 "state": "online", 00:10:39.801 "raid_level": "raid1", 00:10:39.801 "superblock": true, 00:10:39.801 "num_base_bdevs": 3, 00:10:39.801 "num_base_bdevs_discovered": 3, 00:10:39.801 "num_base_bdevs_operational": 3, 00:10:39.801 "base_bdevs_list": [ 00:10:39.801 { 00:10:39.801 "name": "pt1", 00:10:39.801 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:39.801 "is_configured": true, 00:10:39.801 "data_offset": 2048, 00:10:39.801 "data_size": 63488 00:10:39.801 }, 00:10:39.801 { 00:10:39.801 "name": "pt2", 00:10:39.801 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:39.801 "is_configured": true, 00:10:39.801 "data_offset": 2048, 00:10:39.801 "data_size": 63488 00:10:39.801 }, 00:10:39.801 { 00:10:39.801 "name": "pt3", 00:10:39.801 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:39.801 "is_configured": true, 00:10:39.801 "data_offset": 2048, 00:10:39.801 "data_size": 63488 00:10:39.801 } 00:10:39.801 ] 00:10:39.801 }' 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:39.801 15:58:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:40.059 [2024-11-20 15:58:38.189935] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:40.059 "name": "raid_bdev1", 00:10:40.059 "aliases": [ 00:10:40.059 "eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f" 00:10:40.059 ], 00:10:40.059 "product_name": "Raid Volume", 00:10:40.059 "block_size": 512, 00:10:40.059 "num_blocks": 63488, 00:10:40.059 "uuid": "eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f", 00:10:40.059 "assigned_rate_limits": { 00:10:40.059 "rw_ios_per_sec": 0, 00:10:40.059 "rw_mbytes_per_sec": 0, 00:10:40.059 "r_mbytes_per_sec": 0, 00:10:40.059 "w_mbytes_per_sec": 0 00:10:40.059 }, 00:10:40.059 "claimed": false, 00:10:40.059 "zoned": false, 00:10:40.059 "supported_io_types": { 00:10:40.059 "read": true, 00:10:40.059 "write": true, 00:10:40.059 "unmap": false, 00:10:40.059 "flush": false, 00:10:40.059 "reset": true, 00:10:40.059 "nvme_admin": false, 00:10:40.059 "nvme_io": false, 00:10:40.059 "nvme_io_md": false, 00:10:40.059 "write_zeroes": true, 00:10:40.059 "zcopy": false, 00:10:40.059 "get_zone_info": false, 00:10:40.059 "zone_management": false, 00:10:40.059 "zone_append": false, 00:10:40.059 "compare": false, 00:10:40.059 "compare_and_write": false, 00:10:40.059 "abort": false, 00:10:40.059 "seek_hole": false, 00:10:40.059 "seek_data": false, 00:10:40.059 "copy": false, 00:10:40.059 "nvme_iov_md": false 00:10:40.059 }, 00:10:40.059 "memory_domains": [ 00:10:40.059 { 00:10:40.059 "dma_device_id": "system", 00:10:40.059 "dma_device_type": 1 00:10:40.059 }, 00:10:40.059 { 00:10:40.059 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:40.059 "dma_device_type": 2 00:10:40.059 }, 00:10:40.059 { 00:10:40.059 "dma_device_id": "system", 00:10:40.059 "dma_device_type": 1 00:10:40.059 }, 00:10:40.059 { 00:10:40.059 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:40.059 "dma_device_type": 2 00:10:40.059 }, 00:10:40.059 { 00:10:40.059 "dma_device_id": "system", 00:10:40.059 "dma_device_type": 1 00:10:40.059 }, 00:10:40.059 { 00:10:40.059 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:40.059 "dma_device_type": 2 00:10:40.059 } 00:10:40.059 ], 00:10:40.059 "driver_specific": { 00:10:40.059 "raid": { 00:10:40.059 "uuid": "eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f", 00:10:40.059 "strip_size_kb": 0, 00:10:40.059 "state": "online", 00:10:40.059 "raid_level": "raid1", 00:10:40.059 "superblock": true, 00:10:40.059 "num_base_bdevs": 3, 00:10:40.059 "num_base_bdevs_discovered": 3, 00:10:40.059 "num_base_bdevs_operational": 3, 00:10:40.059 "base_bdevs_list": [ 00:10:40.059 { 00:10:40.059 "name": "pt1", 00:10:40.059 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:40.059 "is_configured": true, 00:10:40.059 "data_offset": 2048, 00:10:40.059 "data_size": 63488 00:10:40.059 }, 00:10:40.059 { 00:10:40.059 "name": "pt2", 00:10:40.059 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:40.059 "is_configured": true, 00:10:40.059 "data_offset": 2048, 00:10:40.059 "data_size": 63488 00:10:40.059 }, 00:10:40.059 { 00:10:40.059 "name": "pt3", 00:10:40.059 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:40.059 "is_configured": true, 00:10:40.059 "data_offset": 2048, 00:10:40.059 "data_size": 63488 00:10:40.059 } 00:10:40.059 ] 00:10:40.059 } 00:10:40.059 } 00:10:40.059 }' 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:40.059 pt2 00:10:40.059 pt3' 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.059 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.317 [2024-11-20 15:58:38.382031] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f ']' 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.317 [2024-11-20 15:58:38.405817] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:40.317 [2024-11-20 15:58:38.405839] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:40.317 [2024-11-20 15:58:38.405900] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:40.317 [2024-11-20 15:58:38.405963] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:40.317 [2024-11-20 15:58:38.405971] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.317 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.318 [2024-11-20 15:58:38.509993] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:40.318 [2024-11-20 15:58:38.511524] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:40.318 [2024-11-20 15:58:38.511569] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:40.318 [2024-11-20 15:58:38.511607] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:40.318 [2024-11-20 15:58:38.511654] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:40.318 [2024-11-20 15:58:38.511683] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:40.318 [2024-11-20 15:58:38.511697] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:40.318 [2024-11-20 15:58:38.511705] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:10:40.318 request: 00:10:40.318 { 00:10:40.318 "name": "raid_bdev1", 00:10:40.318 "raid_level": "raid1", 00:10:40.318 "base_bdevs": [ 00:10:40.318 "malloc1", 00:10:40.318 "malloc2", 00:10:40.318 "malloc3" 00:10:40.318 ], 00:10:40.318 "superblock": false, 00:10:40.318 "method": "bdev_raid_create", 00:10:40.318 "req_id": 1 00:10:40.318 } 00:10:40.318 Got JSON-RPC error response 00:10:40.318 response: 00:10:40.318 { 00:10:40.318 "code": -17, 00:10:40.318 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:40.318 } 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.318 [2024-11-20 15:58:38.554002] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:40.318 [2024-11-20 15:58:38.554046] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:40.318 [2024-11-20 15:58:38.554061] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:40.318 [2024-11-20 15:58:38.554068] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:40.318 [2024-11-20 15:58:38.555864] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:40.318 [2024-11-20 15:58:38.555894] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:40.318 [2024-11-20 15:58:38.555959] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:40.318 [2024-11-20 15:58:38.555998] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:40.318 pt1 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:40.318 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.587 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.587 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:40.587 "name": "raid_bdev1", 00:10:40.587 "uuid": "eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f", 00:10:40.587 "strip_size_kb": 0, 00:10:40.587 "state": "configuring", 00:10:40.587 "raid_level": "raid1", 00:10:40.587 "superblock": true, 00:10:40.587 "num_base_bdevs": 3, 00:10:40.587 "num_base_bdevs_discovered": 1, 00:10:40.587 "num_base_bdevs_operational": 3, 00:10:40.587 "base_bdevs_list": [ 00:10:40.587 { 00:10:40.587 "name": "pt1", 00:10:40.587 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:40.587 "is_configured": true, 00:10:40.587 "data_offset": 2048, 00:10:40.587 "data_size": 63488 00:10:40.587 }, 00:10:40.587 { 00:10:40.587 "name": null, 00:10:40.587 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:40.587 "is_configured": false, 00:10:40.587 "data_offset": 2048, 00:10:40.587 "data_size": 63488 00:10:40.587 }, 00:10:40.587 { 00:10:40.587 "name": null, 00:10:40.587 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:40.587 "is_configured": false, 00:10:40.587 "data_offset": 2048, 00:10:40.587 "data_size": 63488 00:10:40.587 } 00:10:40.587 ] 00:10:40.587 }' 00:10:40.587 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:40.587 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.846 [2024-11-20 15:58:38.866374] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:40.846 [2024-11-20 15:58:38.866422] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:40.846 [2024-11-20 15:58:38.866439] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:10:40.846 [2024-11-20 15:58:38.866446] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:40.846 [2024-11-20 15:58:38.866797] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:40.846 [2024-11-20 15:58:38.866809] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:40.846 [2024-11-20 15:58:38.866874] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:40.846 [2024-11-20 15:58:38.866890] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:40.846 pt2 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.846 [2024-11-20 15:58:38.874376] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.846 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:40.846 "name": "raid_bdev1", 00:10:40.846 "uuid": "eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f", 00:10:40.846 "strip_size_kb": 0, 00:10:40.846 "state": "configuring", 00:10:40.846 "raid_level": "raid1", 00:10:40.846 "superblock": true, 00:10:40.847 "num_base_bdevs": 3, 00:10:40.847 "num_base_bdevs_discovered": 1, 00:10:40.847 "num_base_bdevs_operational": 3, 00:10:40.847 "base_bdevs_list": [ 00:10:40.847 { 00:10:40.847 "name": "pt1", 00:10:40.847 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:40.847 "is_configured": true, 00:10:40.847 "data_offset": 2048, 00:10:40.847 "data_size": 63488 00:10:40.847 }, 00:10:40.847 { 00:10:40.847 "name": null, 00:10:40.847 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:40.847 "is_configured": false, 00:10:40.847 "data_offset": 0, 00:10:40.847 "data_size": 63488 00:10:40.847 }, 00:10:40.847 { 00:10:40.847 "name": null, 00:10:40.847 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:40.847 "is_configured": false, 00:10:40.847 "data_offset": 2048, 00:10:40.847 "data_size": 63488 00:10:40.847 } 00:10:40.847 ] 00:10:40.847 }' 00:10:40.847 15:58:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:40.847 15:58:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.115 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:41.115 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:41.115 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:41.115 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.115 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.115 [2024-11-20 15:58:39.182731] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:41.116 [2024-11-20 15:58:39.182786] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:41.116 [2024-11-20 15:58:39.182799] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:10:41.116 [2024-11-20 15:58:39.182808] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:41.116 [2024-11-20 15:58:39.183157] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:41.116 [2024-11-20 15:58:39.183169] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:41.116 [2024-11-20 15:58:39.183227] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:41.116 [2024-11-20 15:58:39.183249] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:41.116 pt2 00:10:41.116 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.116 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:41.116 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:41.116 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:41.116 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.116 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.116 [2024-11-20 15:58:39.190723] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:41.116 [2024-11-20 15:58:39.190762] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:41.116 [2024-11-20 15:58:39.190774] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:41.116 [2024-11-20 15:58:39.190781] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:41.116 [2024-11-20 15:58:39.191090] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:41.116 [2024-11-20 15:58:39.191104] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:41.116 [2024-11-20 15:58:39.191154] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:41.116 [2024-11-20 15:58:39.191169] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:41.116 [2024-11-20 15:58:39.191266] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:10:41.116 [2024-11-20 15:58:39.191276] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:41.116 [2024-11-20 15:58:39.191467] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:10:41.116 [2024-11-20 15:58:39.191575] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:10:41.117 [2024-11-20 15:58:39.191581] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:10:41.117 [2024-11-20 15:58:39.191696] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:41.117 pt3 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.117 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:41.117 "name": "raid_bdev1", 00:10:41.117 "uuid": "eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f", 00:10:41.117 "strip_size_kb": 0, 00:10:41.117 "state": "online", 00:10:41.117 "raid_level": "raid1", 00:10:41.117 "superblock": true, 00:10:41.117 "num_base_bdevs": 3, 00:10:41.117 "num_base_bdevs_discovered": 3, 00:10:41.117 "num_base_bdevs_operational": 3, 00:10:41.117 "base_bdevs_list": [ 00:10:41.118 { 00:10:41.118 "name": "pt1", 00:10:41.118 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:41.118 "is_configured": true, 00:10:41.118 "data_offset": 2048, 00:10:41.118 "data_size": 63488 00:10:41.118 }, 00:10:41.118 { 00:10:41.118 "name": "pt2", 00:10:41.118 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:41.118 "is_configured": true, 00:10:41.118 "data_offset": 2048, 00:10:41.118 "data_size": 63488 00:10:41.118 }, 00:10:41.118 { 00:10:41.118 "name": "pt3", 00:10:41.118 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:41.119 "is_configured": true, 00:10:41.119 "data_offset": 2048, 00:10:41.119 "data_size": 63488 00:10:41.119 } 00:10:41.119 ] 00:10:41.119 }' 00:10:41.119 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:41.119 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.379 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:41.379 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:41.379 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:41.379 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:41.379 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:41.379 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:41.379 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:41.379 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:41.379 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.379 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.379 [2024-11-20 15:58:39.507365] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:41.379 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.379 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:41.379 "name": "raid_bdev1", 00:10:41.379 "aliases": [ 00:10:41.379 "eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f" 00:10:41.379 ], 00:10:41.379 "product_name": "Raid Volume", 00:10:41.379 "block_size": 512, 00:10:41.379 "num_blocks": 63488, 00:10:41.379 "uuid": "eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f", 00:10:41.379 "assigned_rate_limits": { 00:10:41.379 "rw_ios_per_sec": 0, 00:10:41.379 "rw_mbytes_per_sec": 0, 00:10:41.379 "r_mbytes_per_sec": 0, 00:10:41.379 "w_mbytes_per_sec": 0 00:10:41.379 }, 00:10:41.379 "claimed": false, 00:10:41.379 "zoned": false, 00:10:41.379 "supported_io_types": { 00:10:41.379 "read": true, 00:10:41.379 "write": true, 00:10:41.379 "unmap": false, 00:10:41.379 "flush": false, 00:10:41.379 "reset": true, 00:10:41.379 "nvme_admin": false, 00:10:41.379 "nvme_io": false, 00:10:41.379 "nvme_io_md": false, 00:10:41.379 "write_zeroes": true, 00:10:41.379 "zcopy": false, 00:10:41.379 "get_zone_info": false, 00:10:41.379 "zone_management": false, 00:10:41.379 "zone_append": false, 00:10:41.379 "compare": false, 00:10:41.379 "compare_and_write": false, 00:10:41.379 "abort": false, 00:10:41.379 "seek_hole": false, 00:10:41.379 "seek_data": false, 00:10:41.379 "copy": false, 00:10:41.379 "nvme_iov_md": false 00:10:41.379 }, 00:10:41.379 "memory_domains": [ 00:10:41.379 { 00:10:41.379 "dma_device_id": "system", 00:10:41.379 "dma_device_type": 1 00:10:41.379 }, 00:10:41.379 { 00:10:41.379 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:41.379 "dma_device_type": 2 00:10:41.379 }, 00:10:41.379 { 00:10:41.379 "dma_device_id": "system", 00:10:41.379 "dma_device_type": 1 00:10:41.379 }, 00:10:41.379 { 00:10:41.379 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:41.380 "dma_device_type": 2 00:10:41.380 }, 00:10:41.380 { 00:10:41.380 "dma_device_id": "system", 00:10:41.380 "dma_device_type": 1 00:10:41.380 }, 00:10:41.380 { 00:10:41.380 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:41.380 "dma_device_type": 2 00:10:41.380 } 00:10:41.380 ], 00:10:41.380 "driver_specific": { 00:10:41.380 "raid": { 00:10:41.380 "uuid": "eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f", 00:10:41.380 "strip_size_kb": 0, 00:10:41.380 "state": "online", 00:10:41.380 "raid_level": "raid1", 00:10:41.380 "superblock": true, 00:10:41.380 "num_base_bdevs": 3, 00:10:41.380 "num_base_bdevs_discovered": 3, 00:10:41.380 "num_base_bdevs_operational": 3, 00:10:41.380 "base_bdevs_list": [ 00:10:41.380 { 00:10:41.380 "name": "pt1", 00:10:41.380 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:41.380 "is_configured": true, 00:10:41.380 "data_offset": 2048, 00:10:41.380 "data_size": 63488 00:10:41.380 }, 00:10:41.380 { 00:10:41.380 "name": "pt2", 00:10:41.380 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:41.380 "is_configured": true, 00:10:41.380 "data_offset": 2048, 00:10:41.380 "data_size": 63488 00:10:41.380 }, 00:10:41.380 { 00:10:41.380 "name": "pt3", 00:10:41.380 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:41.380 "is_configured": true, 00:10:41.380 "data_offset": 2048, 00:10:41.380 "data_size": 63488 00:10:41.380 } 00:10:41.380 ] 00:10:41.380 } 00:10:41.380 } 00:10:41.380 }' 00:10:41.380 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:41.380 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:41.380 pt2 00:10:41.380 pt3' 00:10:41.380 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:41.380 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:41.380 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:41.380 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:41.380 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:41.380 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.380 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.380 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.380 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:41.380 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:41.380 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:41.380 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:41.380 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.380 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.638 [2024-11-20 15:58:39.703548] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f '!=' eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f ']' 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.638 [2024-11-20 15:58:39.731361] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:41.638 "name": "raid_bdev1", 00:10:41.638 "uuid": "eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f", 00:10:41.638 "strip_size_kb": 0, 00:10:41.638 "state": "online", 00:10:41.638 "raid_level": "raid1", 00:10:41.638 "superblock": true, 00:10:41.638 "num_base_bdevs": 3, 00:10:41.638 "num_base_bdevs_discovered": 2, 00:10:41.638 "num_base_bdevs_operational": 2, 00:10:41.638 "base_bdevs_list": [ 00:10:41.638 { 00:10:41.638 "name": null, 00:10:41.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:41.638 "is_configured": false, 00:10:41.638 "data_offset": 0, 00:10:41.638 "data_size": 63488 00:10:41.638 }, 00:10:41.638 { 00:10:41.638 "name": "pt2", 00:10:41.638 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:41.638 "is_configured": true, 00:10:41.638 "data_offset": 2048, 00:10:41.638 "data_size": 63488 00:10:41.638 }, 00:10:41.638 { 00:10:41.638 "name": "pt3", 00:10:41.638 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:41.638 "is_configured": true, 00:10:41.638 "data_offset": 2048, 00:10:41.638 "data_size": 63488 00:10:41.638 } 00:10:41.638 ] 00:10:41.638 }' 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:41.638 15:58:39 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.896 [2024-11-20 15:58:40.031650] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:41.896 [2024-11-20 15:58:40.031681] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:41.896 [2024-11-20 15:58:40.031739] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:41.896 [2024-11-20 15:58:40.031786] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:41.896 [2024-11-20 15:58:40.031797] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.896 [2024-11-20 15:58:40.091712] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:41.896 [2024-11-20 15:58:40.091757] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:41.896 [2024-11-20 15:58:40.091772] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:10:41.896 [2024-11-20 15:58:40.091780] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:41.896 [2024-11-20 15:58:40.093612] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:41.896 [2024-11-20 15:58:40.093648] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:41.896 [2024-11-20 15:58:40.093725] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:41.896 [2024-11-20 15:58:40.093763] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:41.896 pt2 00:10:41.896 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:41.897 "name": "raid_bdev1", 00:10:41.897 "uuid": "eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f", 00:10:41.897 "strip_size_kb": 0, 00:10:41.897 "state": "configuring", 00:10:41.897 "raid_level": "raid1", 00:10:41.897 "superblock": true, 00:10:41.897 "num_base_bdevs": 3, 00:10:41.897 "num_base_bdevs_discovered": 1, 00:10:41.897 "num_base_bdevs_operational": 2, 00:10:41.897 "base_bdevs_list": [ 00:10:41.897 { 00:10:41.897 "name": null, 00:10:41.897 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:41.897 "is_configured": false, 00:10:41.897 "data_offset": 2048, 00:10:41.897 "data_size": 63488 00:10:41.897 }, 00:10:41.897 { 00:10:41.897 "name": "pt2", 00:10:41.897 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:41.897 "is_configured": true, 00:10:41.897 "data_offset": 2048, 00:10:41.897 "data_size": 63488 00:10:41.897 }, 00:10:41.897 { 00:10:41.897 "name": null, 00:10:41.897 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:41.897 "is_configured": false, 00:10:41.897 "data_offset": 2048, 00:10:41.897 "data_size": 63488 00:10:41.897 } 00:10:41.897 ] 00:10:41.897 }' 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:41.897 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.464 [2024-11-20 15:58:40.436100] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:42.464 [2024-11-20 15:58:40.436152] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:42.464 [2024-11-20 15:58:40.436167] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:10:42.464 [2024-11-20 15:58:40.436176] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:42.464 [2024-11-20 15:58:40.436528] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:42.464 [2024-11-20 15:58:40.436541] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:42.464 [2024-11-20 15:58:40.436603] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:42.464 [2024-11-20 15:58:40.436623] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:42.464 [2024-11-20 15:58:40.436730] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:10:42.464 [2024-11-20 15:58:40.436739] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:42.464 [2024-11-20 15:58:40.436941] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:10:42.464 [2024-11-20 15:58:40.437050] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:10:42.464 [2024-11-20 15:58:40.437061] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:10:42.464 [2024-11-20 15:58:40.437166] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:42.464 pt3 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.464 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:42.464 "name": "raid_bdev1", 00:10:42.464 "uuid": "eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f", 00:10:42.464 "strip_size_kb": 0, 00:10:42.464 "state": "online", 00:10:42.464 "raid_level": "raid1", 00:10:42.464 "superblock": true, 00:10:42.464 "num_base_bdevs": 3, 00:10:42.464 "num_base_bdevs_discovered": 2, 00:10:42.464 "num_base_bdevs_operational": 2, 00:10:42.464 "base_bdevs_list": [ 00:10:42.464 { 00:10:42.464 "name": null, 00:10:42.464 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:42.464 "is_configured": false, 00:10:42.464 "data_offset": 2048, 00:10:42.464 "data_size": 63488 00:10:42.464 }, 00:10:42.464 { 00:10:42.465 "name": "pt2", 00:10:42.465 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:42.465 "is_configured": true, 00:10:42.465 "data_offset": 2048, 00:10:42.465 "data_size": 63488 00:10:42.465 }, 00:10:42.465 { 00:10:42.465 "name": "pt3", 00:10:42.465 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:42.465 "is_configured": true, 00:10:42.465 "data_offset": 2048, 00:10:42.465 "data_size": 63488 00:10:42.465 } 00:10:42.465 ] 00:10:42.465 }' 00:10:42.465 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:42.465 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.723 [2024-11-20 15:58:40.768438] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:42.723 [2024-11-20 15:58:40.768459] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:42.723 [2024-11-20 15:58:40.768516] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:42.723 [2024-11-20 15:58:40.768565] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:42.723 [2024-11-20 15:58:40.768572] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.723 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.723 [2024-11-20 15:58:40.820512] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:42.723 [2024-11-20 15:58:40.820555] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:42.723 [2024-11-20 15:58:40.820569] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:10:42.723 [2024-11-20 15:58:40.820577] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:42.724 [2024-11-20 15:58:40.822419] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:42.724 [2024-11-20 15:58:40.822445] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:42.724 [2024-11-20 15:58:40.822508] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:42.724 [2024-11-20 15:58:40.822541] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:42.724 [2024-11-20 15:58:40.822635] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:10:42.724 [2024-11-20 15:58:40.822644] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:42.724 [2024-11-20 15:58:40.822657] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:10:42.724 [2024-11-20 15:58:40.822708] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:42.724 pt1 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:42.724 "name": "raid_bdev1", 00:10:42.724 "uuid": "eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f", 00:10:42.724 "strip_size_kb": 0, 00:10:42.724 "state": "configuring", 00:10:42.724 "raid_level": "raid1", 00:10:42.724 "superblock": true, 00:10:42.724 "num_base_bdevs": 3, 00:10:42.724 "num_base_bdevs_discovered": 1, 00:10:42.724 "num_base_bdevs_operational": 2, 00:10:42.724 "base_bdevs_list": [ 00:10:42.724 { 00:10:42.724 "name": null, 00:10:42.724 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:42.724 "is_configured": false, 00:10:42.724 "data_offset": 2048, 00:10:42.724 "data_size": 63488 00:10:42.724 }, 00:10:42.724 { 00:10:42.724 "name": "pt2", 00:10:42.724 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:42.724 "is_configured": true, 00:10:42.724 "data_offset": 2048, 00:10:42.724 "data_size": 63488 00:10:42.724 }, 00:10:42.724 { 00:10:42.724 "name": null, 00:10:42.724 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:42.724 "is_configured": false, 00:10:42.724 "data_offset": 2048, 00:10:42.724 "data_size": 63488 00:10:42.724 } 00:10:42.724 ] 00:10:42.724 }' 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:42.724 15:58:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.981 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:10:42.981 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:10:42.981 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.981 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.981 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.981 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.982 [2024-11-20 15:58:41.212952] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:42.982 [2024-11-20 15:58:41.213105] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:42.982 [2024-11-20 15:58:41.213125] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:10:42.982 [2024-11-20 15:58:41.213132] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:42.982 [2024-11-20 15:58:41.213496] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:42.982 [2024-11-20 15:58:41.213508] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:42.982 [2024-11-20 15:58:41.213569] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:42.982 [2024-11-20 15:58:41.213584] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:42.982 [2024-11-20 15:58:41.213690] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:10:42.982 [2024-11-20 15:58:41.213698] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:42.982 [2024-11-20 15:58:41.213897] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:10:42.982 [2024-11-20 15:58:41.214005] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:10:42.982 [2024-11-20 15:58:41.214020] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:10:42.982 [2024-11-20 15:58:41.214125] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:42.982 pt3 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.982 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.239 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:43.239 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:43.239 "name": "raid_bdev1", 00:10:43.239 "uuid": "eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f", 00:10:43.239 "strip_size_kb": 0, 00:10:43.239 "state": "online", 00:10:43.239 "raid_level": "raid1", 00:10:43.239 "superblock": true, 00:10:43.239 "num_base_bdevs": 3, 00:10:43.239 "num_base_bdevs_discovered": 2, 00:10:43.239 "num_base_bdevs_operational": 2, 00:10:43.239 "base_bdevs_list": [ 00:10:43.239 { 00:10:43.239 "name": null, 00:10:43.239 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:43.239 "is_configured": false, 00:10:43.239 "data_offset": 2048, 00:10:43.239 "data_size": 63488 00:10:43.239 }, 00:10:43.239 { 00:10:43.239 "name": "pt2", 00:10:43.239 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:43.239 "is_configured": true, 00:10:43.239 "data_offset": 2048, 00:10:43.239 "data_size": 63488 00:10:43.239 }, 00:10:43.239 { 00:10:43.239 "name": "pt3", 00:10:43.239 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:43.239 "is_configured": true, 00:10:43.239 "data_offset": 2048, 00:10:43.239 "data_size": 63488 00:10:43.239 } 00:10:43.239 ] 00:10:43.239 }' 00:10:43.239 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:43.239 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.495 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:10:43.496 [2024-11-20 15:58:41.585553] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f '!=' eaec5e1a-4acb-495d-ab0d-51cc4b87fe4f ']' 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 66883 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 66883 ']' 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 66883 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 66883 00:10:43.496 killing process with pid 66883 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 66883' 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 66883 00:10:43.496 [2024-11-20 15:58:41.633329] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:43.496 [2024-11-20 15:58:41.633408] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:43.496 15:58:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 66883 00:10:43.496 [2024-11-20 15:58:41.633457] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:43.496 [2024-11-20 15:58:41.633466] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:10:43.754 [2024-11-20 15:58:41.780704] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:44.319 15:58:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:44.319 00:10:44.319 real 0m5.480s 00:10:44.319 user 0m8.728s 00:10:44.319 sys 0m0.866s 00:10:44.319 ************************************ 00:10:44.319 END TEST raid_superblock_test 00:10:44.319 ************************************ 00:10:44.319 15:58:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:44.319 15:58:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.319 15:58:42 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 3 read 00:10:44.319 15:58:42 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:44.319 15:58:42 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:44.319 15:58:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:44.319 ************************************ 00:10:44.319 START TEST raid_read_error_test 00:10:44.319 ************************************ 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 3 read 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.csOaxSBeaE 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=67301 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 67301 00:10:44.319 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 67301 ']' 00:10:44.319 15:58:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:44.320 15:58:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:44.320 15:58:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:44.320 15:58:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:44.320 15:58:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:44.320 15:58:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.320 [2024-11-20 15:58:42.471748] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:10:44.320 [2024-11-20 15:58:42.471869] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid67301 ] 00:10:44.577 [2024-11-20 15:58:42.628347] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:44.577 [2024-11-20 15:58:42.712085] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:44.577 [2024-11-20 15:58:42.821044] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:44.577 [2024-11-20 15:58:42.821080] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.143 BaseBdev1_malloc 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.143 true 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.143 [2024-11-20 15:58:43.343312] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:45.143 [2024-11-20 15:58:43.343496] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:45.143 [2024-11-20 15:58:43.343517] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:10:45.143 [2024-11-20 15:58:43.343526] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:45.143 [2024-11-20 15:58:43.345264] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:45.143 [2024-11-20 15:58:43.345295] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:45.143 BaseBdev1 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.143 BaseBdev2_malloc 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.143 true 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.143 [2024-11-20 15:58:43.382419] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:45.143 [2024-11-20 15:58:43.382548] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:45.143 [2024-11-20 15:58:43.382565] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:45.143 [2024-11-20 15:58:43.382573] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:45.143 [2024-11-20 15:58:43.384333] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:45.143 [2024-11-20 15:58:43.384359] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:45.143 BaseBdev2 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.143 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.402 BaseBdev3_malloc 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.402 true 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.402 [2024-11-20 15:58:43.436857] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:45.402 [2024-11-20 15:58:43.436904] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:45.402 [2024-11-20 15:58:43.436918] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:45.402 [2024-11-20 15:58:43.436926] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:45.402 [2024-11-20 15:58:43.438630] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:45.402 [2024-11-20 15:58:43.438774] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:45.402 BaseBdev3 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.402 [2024-11-20 15:58:43.444913] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:45.402 [2024-11-20 15:58:43.446395] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:45.402 [2024-11-20 15:58:43.446450] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:45.402 [2024-11-20 15:58:43.446608] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:10:45.402 [2024-11-20 15:58:43.446617] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:45.402 [2024-11-20 15:58:43.446915] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:10:45.402 [2024-11-20 15:58:43.447111] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:10:45.402 [2024-11-20 15:58:43.447212] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:10:45.402 [2024-11-20 15:58:43.447375] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.402 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:45.402 "name": "raid_bdev1", 00:10:45.402 "uuid": "bba1eb51-db60-473f-baaf-0c30e0eba6ff", 00:10:45.402 "strip_size_kb": 0, 00:10:45.402 "state": "online", 00:10:45.402 "raid_level": "raid1", 00:10:45.402 "superblock": true, 00:10:45.402 "num_base_bdevs": 3, 00:10:45.402 "num_base_bdevs_discovered": 3, 00:10:45.403 "num_base_bdevs_operational": 3, 00:10:45.403 "base_bdevs_list": [ 00:10:45.403 { 00:10:45.403 "name": "BaseBdev1", 00:10:45.403 "uuid": "8d9f0d3c-8e86-5890-afc6-8f3c003981ea", 00:10:45.403 "is_configured": true, 00:10:45.403 "data_offset": 2048, 00:10:45.403 "data_size": 63488 00:10:45.403 }, 00:10:45.403 { 00:10:45.403 "name": "BaseBdev2", 00:10:45.403 "uuid": "49451f95-9699-5807-ba92-539ca3140730", 00:10:45.403 "is_configured": true, 00:10:45.403 "data_offset": 2048, 00:10:45.403 "data_size": 63488 00:10:45.403 }, 00:10:45.403 { 00:10:45.403 "name": "BaseBdev3", 00:10:45.403 "uuid": "f0a1caa8-5de5-56c7-8e0b-daf8e5e65980", 00:10:45.403 "is_configured": true, 00:10:45.403 "data_offset": 2048, 00:10:45.403 "data_size": 63488 00:10:45.403 } 00:10:45.403 ] 00:10:45.403 }' 00:10:45.403 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:45.403 15:58:43 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.660 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:45.660 15:58:43 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:45.660 [2024-11-20 15:58:43.833739] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:46.592 "name": "raid_bdev1", 00:10:46.592 "uuid": "bba1eb51-db60-473f-baaf-0c30e0eba6ff", 00:10:46.592 "strip_size_kb": 0, 00:10:46.592 "state": "online", 00:10:46.592 "raid_level": "raid1", 00:10:46.592 "superblock": true, 00:10:46.592 "num_base_bdevs": 3, 00:10:46.592 "num_base_bdevs_discovered": 3, 00:10:46.592 "num_base_bdevs_operational": 3, 00:10:46.592 "base_bdevs_list": [ 00:10:46.592 { 00:10:46.592 "name": "BaseBdev1", 00:10:46.592 "uuid": "8d9f0d3c-8e86-5890-afc6-8f3c003981ea", 00:10:46.592 "is_configured": true, 00:10:46.592 "data_offset": 2048, 00:10:46.592 "data_size": 63488 00:10:46.592 }, 00:10:46.592 { 00:10:46.592 "name": "BaseBdev2", 00:10:46.592 "uuid": "49451f95-9699-5807-ba92-539ca3140730", 00:10:46.592 "is_configured": true, 00:10:46.592 "data_offset": 2048, 00:10:46.592 "data_size": 63488 00:10:46.592 }, 00:10:46.592 { 00:10:46.592 "name": "BaseBdev3", 00:10:46.592 "uuid": "f0a1caa8-5de5-56c7-8e0b-daf8e5e65980", 00:10:46.592 "is_configured": true, 00:10:46.592 "data_offset": 2048, 00:10:46.592 "data_size": 63488 00:10:46.592 } 00:10:46.592 ] 00:10:46.592 }' 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:46.592 15:58:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.850 15:58:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:46.850 15:58:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.850 15:58:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.850 [2024-11-20 15:58:45.079167] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:46.850 [2024-11-20 15:58:45.079192] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:46.850 [2024-11-20 15:58:45.081688] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:46.850 [2024-11-20 15:58:45.081728] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:46.850 [2024-11-20 15:58:45.081817] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:46.850 [2024-11-20 15:58:45.081825] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:10:46.850 { 00:10:46.850 "results": [ 00:10:46.850 { 00:10:46.850 "job": "raid_bdev1", 00:10:46.850 "core_mask": "0x1", 00:10:46.850 "workload": "randrw", 00:10:46.850 "percentage": 50, 00:10:46.850 "status": "finished", 00:10:46.850 "queue_depth": 1, 00:10:46.850 "io_size": 131072, 00:10:46.850 "runtime": 1.24388, 00:10:46.850 "iops": 17236.38936231791, 00:10:46.850 "mibps": 2154.5486702897388, 00:10:46.850 "io_failed": 0, 00:10:46.850 "io_timeout": 0, 00:10:46.850 "avg_latency_us": 55.47757520091849, 00:10:46.850 "min_latency_us": 23.433846153846154, 00:10:46.850 "max_latency_us": 1329.6246153846155 00:10:46.850 } 00:10:46.850 ], 00:10:46.850 "core_count": 1 00:10:46.850 } 00:10:46.850 15:58:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.850 15:58:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 67301 00:10:46.850 15:58:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 67301 ']' 00:10:46.850 15:58:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 67301 00:10:46.850 15:58:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:10:46.850 15:58:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:46.850 15:58:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 67301 00:10:47.107 killing process with pid 67301 00:10:47.107 15:58:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:47.107 15:58:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:47.107 15:58:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 67301' 00:10:47.107 15:58:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 67301 00:10:47.107 [2024-11-20 15:58:45.110557] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:47.107 15:58:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 67301 00:10:47.107 [2024-11-20 15:58:45.222337] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:47.671 15:58:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.csOaxSBeaE 00:10:47.671 15:58:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:47.671 15:58:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:47.671 15:58:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:10:47.671 15:58:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:10:47.671 15:58:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:47.671 15:58:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:47.671 15:58:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:10:47.671 00:10:47.671 real 0m3.430s 00:10:47.671 user 0m4.105s 00:10:47.671 sys 0m0.381s 00:10:47.671 15:58:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:47.671 15:58:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.671 ************************************ 00:10:47.671 END TEST raid_read_error_test 00:10:47.671 ************************************ 00:10:47.671 15:58:45 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 3 write 00:10:47.671 15:58:45 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:47.671 15:58:45 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:47.671 15:58:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:47.671 ************************************ 00:10:47.671 START TEST raid_write_error_test 00:10:47.671 ************************************ 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 3 write 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:47.671 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:47.672 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:47.672 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:47.672 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:47.672 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:47.672 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:10:47.672 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:10:47.672 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:47.672 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.3vsPSS30nu 00:10:47.672 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:47.672 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=67430 00:10:47.672 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 67430 00:10:47.672 15:58:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 67430 ']' 00:10:47.672 15:58:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:47.672 15:58:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:47.672 15:58:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:47.672 15:58:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:47.672 15:58:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.672 15:58:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:47.929 [2024-11-20 15:58:45.937040] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:10:47.929 [2024-11-20 15:58:45.937154] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid67430 ] 00:10:47.929 [2024-11-20 15:58:46.091967] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:47.929 [2024-11-20 15:58:46.174469] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:48.185 [2024-11-20 15:58:46.282384] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:48.185 [2024-11-20 15:58:46.282573] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.781 BaseBdev1_malloc 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.781 true 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.781 [2024-11-20 15:58:46.770890] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:48.781 [2024-11-20 15:58:46.771020] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:48.781 [2024-11-20 15:58:46.771053] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:10:48.781 [2024-11-20 15:58:46.771107] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:48.781 [2024-11-20 15:58:46.772790] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:48.781 [2024-11-20 15:58:46.772890] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:48.781 BaseBdev1 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.781 BaseBdev2_malloc 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.781 true 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.781 [2024-11-20 15:58:46.813542] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:48.781 [2024-11-20 15:58:46.813653] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:48.781 [2024-11-20 15:58:46.813731] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:48.781 [2024-11-20 15:58:46.813803] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:48.781 [2024-11-20 15:58:46.815485] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:48.781 [2024-11-20 15:58:46.815571] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:48.781 BaseBdev2 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.781 BaseBdev3_malloc 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.781 true 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.781 [2024-11-20 15:58:46.874656] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:48.781 [2024-11-20 15:58:46.874781] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:48.781 [2024-11-20 15:58:46.874811] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:48.781 [2024-11-20 15:58:46.874903] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:48.781 [2024-11-20 15:58:46.876561] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:48.781 [2024-11-20 15:58:46.876591] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:48.781 BaseBdev3 00:10:48.781 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.782 [2024-11-20 15:58:46.882723] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:48.782 [2024-11-20 15:58:46.884223] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:48.782 [2024-11-20 15:58:46.884349] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:48.782 [2024-11-20 15:58:46.884524] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:10:48.782 [2024-11-20 15:58:46.884549] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:48.782 [2024-11-20 15:58:46.884800] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:10:48.782 [2024-11-20 15:58:46.884982] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:10:48.782 [2024-11-20 15:58:46.885041] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:10:48.782 [2024-11-20 15:58:46.885226] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:48.782 "name": "raid_bdev1", 00:10:48.782 "uuid": "aea2a53d-6da0-42e4-89f6-75099efe5c67", 00:10:48.782 "strip_size_kb": 0, 00:10:48.782 "state": "online", 00:10:48.782 "raid_level": "raid1", 00:10:48.782 "superblock": true, 00:10:48.782 "num_base_bdevs": 3, 00:10:48.782 "num_base_bdevs_discovered": 3, 00:10:48.782 "num_base_bdevs_operational": 3, 00:10:48.782 "base_bdevs_list": [ 00:10:48.782 { 00:10:48.782 "name": "BaseBdev1", 00:10:48.782 "uuid": "6981dc25-7232-5d34-a273-a31f6dd3f643", 00:10:48.782 "is_configured": true, 00:10:48.782 "data_offset": 2048, 00:10:48.782 "data_size": 63488 00:10:48.782 }, 00:10:48.782 { 00:10:48.782 "name": "BaseBdev2", 00:10:48.782 "uuid": "dce14ba1-b17c-541d-aa91-ac482274ed61", 00:10:48.782 "is_configured": true, 00:10:48.782 "data_offset": 2048, 00:10:48.782 "data_size": 63488 00:10:48.782 }, 00:10:48.782 { 00:10:48.782 "name": "BaseBdev3", 00:10:48.782 "uuid": "8351bd73-9c7a-58f4-a12b-3534206e9289", 00:10:48.782 "is_configured": true, 00:10:48.782 "data_offset": 2048, 00:10:48.782 "data_size": 63488 00:10:48.782 } 00:10:48.782 ] 00:10:48.782 }' 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:48.782 15:58:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.038 15:58:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:49.038 15:58:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:49.039 [2024-11-20 15:58:47.275538] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.969 [2024-11-20 15:58:48.198059] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:10:49.969 [2024-11-20 15:58:48.198207] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:49.969 [2024-11-20 15:58:48.198406] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006700 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=2 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.969 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.227 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.227 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:50.227 "name": "raid_bdev1", 00:10:50.227 "uuid": "aea2a53d-6da0-42e4-89f6-75099efe5c67", 00:10:50.227 "strip_size_kb": 0, 00:10:50.227 "state": "online", 00:10:50.227 "raid_level": "raid1", 00:10:50.227 "superblock": true, 00:10:50.227 "num_base_bdevs": 3, 00:10:50.227 "num_base_bdevs_discovered": 2, 00:10:50.227 "num_base_bdevs_operational": 2, 00:10:50.227 "base_bdevs_list": [ 00:10:50.227 { 00:10:50.227 "name": null, 00:10:50.227 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:50.227 "is_configured": false, 00:10:50.227 "data_offset": 0, 00:10:50.227 "data_size": 63488 00:10:50.227 }, 00:10:50.227 { 00:10:50.227 "name": "BaseBdev2", 00:10:50.227 "uuid": "dce14ba1-b17c-541d-aa91-ac482274ed61", 00:10:50.227 "is_configured": true, 00:10:50.227 "data_offset": 2048, 00:10:50.227 "data_size": 63488 00:10:50.227 }, 00:10:50.227 { 00:10:50.227 "name": "BaseBdev3", 00:10:50.227 "uuid": "8351bd73-9c7a-58f4-a12b-3534206e9289", 00:10:50.227 "is_configured": true, 00:10:50.227 "data_offset": 2048, 00:10:50.227 "data_size": 63488 00:10:50.227 } 00:10:50.227 ] 00:10:50.227 }' 00:10:50.227 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:50.227 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.485 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:50.485 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.485 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.485 [2024-11-20 15:58:48.508034] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:50.485 [2024-11-20 15:58:48.508143] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:50.485 [2024-11-20 15:58:48.510533] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:50.485 [2024-11-20 15:58:48.510570] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:50.485 [2024-11-20 15:58:48.510637] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:50.485 [2024-11-20 15:58:48.510648] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:10:50.485 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.485 { 00:10:50.485 "results": [ 00:10:50.485 { 00:10:50.485 "job": "raid_bdev1", 00:10:50.485 "core_mask": "0x1", 00:10:50.485 "workload": "randrw", 00:10:50.485 "percentage": 50, 00:10:50.485 "status": "finished", 00:10:50.485 "queue_depth": 1, 00:10:50.485 "io_size": 131072, 00:10:50.485 "runtime": 1.231153, 00:10:50.485 "iops": 18982.20611085706, 00:10:50.485 "mibps": 2372.7757638571325, 00:10:50.485 "io_failed": 0, 00:10:50.485 "io_timeout": 0, 00:10:50.485 "avg_latency_us": 50.19162239557618, 00:10:50.485 "min_latency_us": 23.04, 00:10:50.485 "max_latency_us": 1329.6246153846155 00:10:50.485 } 00:10:50.485 ], 00:10:50.485 "core_count": 1 00:10:50.485 } 00:10:50.485 15:58:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 67430 00:10:50.485 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 67430 ']' 00:10:50.485 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 67430 00:10:50.485 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:10:50.485 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:50.485 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 67430 00:10:50.485 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:50.485 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:50.485 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 67430' 00:10:50.485 killing process with pid 67430 00:10:50.485 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 67430 00:10:50.485 [2024-11-20 15:58:48.537632] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:50.485 15:58:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 67430 00:10:50.485 [2024-11-20 15:58:48.647899] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:51.053 15:58:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.3vsPSS30nu 00:10:51.053 15:58:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:51.053 15:58:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:51.053 15:58:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:10:51.053 ************************************ 00:10:51.053 END TEST raid_write_error_test 00:10:51.053 ************************************ 00:10:51.053 15:58:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:10:51.053 15:58:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:51.053 15:58:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:51.053 15:58:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:10:51.053 00:10:51.053 real 0m3.375s 00:10:51.053 user 0m4.034s 00:10:51.053 sys 0m0.349s 00:10:51.053 15:58:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:51.053 15:58:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.053 15:58:49 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:10:51.053 15:58:49 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:10:51.053 15:58:49 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 4 false 00:10:51.053 15:58:49 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:51.053 15:58:49 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:51.053 15:58:49 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:51.313 ************************************ 00:10:51.313 START TEST raid_state_function_test 00:10:51.313 ************************************ 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 4 false 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:10:51.313 Process raid pid: 67567 00:10:51.313 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=67567 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 67567' 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 67567 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 67567 ']' 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:51.313 15:58:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.313 [2024-11-20 15:58:49.371689] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:10:51.313 [2024-11-20 15:58:49.371928] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:51.313 [2024-11-20 15:58:49.536935] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:51.571 [2024-11-20 15:58:49.620193] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:51.571 [2024-11-20 15:58:49.730370] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:51.571 [2024-11-20 15:58:49.730400] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.136 [2024-11-20 15:58:50.171389] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:52.136 [2024-11-20 15:58:50.171522] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:52.136 [2024-11-20 15:58:50.171575] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:52.136 [2024-11-20 15:58:50.171596] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:52.136 [2024-11-20 15:58:50.171611] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:52.136 [2024-11-20 15:58:50.171627] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:52.136 [2024-11-20 15:58:50.171640] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:52.136 [2024-11-20 15:58:50.171656] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:52.136 "name": "Existed_Raid", 00:10:52.136 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.136 "strip_size_kb": 64, 00:10:52.136 "state": "configuring", 00:10:52.136 "raid_level": "raid0", 00:10:52.136 "superblock": false, 00:10:52.136 "num_base_bdevs": 4, 00:10:52.136 "num_base_bdevs_discovered": 0, 00:10:52.136 "num_base_bdevs_operational": 4, 00:10:52.136 "base_bdevs_list": [ 00:10:52.136 { 00:10:52.136 "name": "BaseBdev1", 00:10:52.136 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.136 "is_configured": false, 00:10:52.136 "data_offset": 0, 00:10:52.136 "data_size": 0 00:10:52.136 }, 00:10:52.136 { 00:10:52.136 "name": "BaseBdev2", 00:10:52.136 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.136 "is_configured": false, 00:10:52.136 "data_offset": 0, 00:10:52.136 "data_size": 0 00:10:52.136 }, 00:10:52.136 { 00:10:52.136 "name": "BaseBdev3", 00:10:52.136 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.136 "is_configured": false, 00:10:52.136 "data_offset": 0, 00:10:52.136 "data_size": 0 00:10:52.136 }, 00:10:52.136 { 00:10:52.136 "name": "BaseBdev4", 00:10:52.136 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.136 "is_configured": false, 00:10:52.136 "data_offset": 0, 00:10:52.136 "data_size": 0 00:10:52.136 } 00:10:52.136 ] 00:10:52.136 }' 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:52.136 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.394 [2024-11-20 15:58:50.491405] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:52.394 [2024-11-20 15:58:50.491516] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.394 [2024-11-20 15:58:50.499423] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:52.394 [2024-11-20 15:58:50.499523] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:52.394 [2024-11-20 15:58:50.499569] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:52.394 [2024-11-20 15:58:50.499591] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:52.394 [2024-11-20 15:58:50.499605] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:52.394 [2024-11-20 15:58:50.499621] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:52.394 [2024-11-20 15:58:50.499634] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:52.394 [2024-11-20 15:58:50.499695] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.394 [2024-11-20 15:58:50.527139] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:52.394 BaseBdev1 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.394 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.394 [ 00:10:52.395 { 00:10:52.395 "name": "BaseBdev1", 00:10:52.395 "aliases": [ 00:10:52.395 "37d0de1d-c2cd-46d3-a5b5-06f0fd623504" 00:10:52.395 ], 00:10:52.395 "product_name": "Malloc disk", 00:10:52.395 "block_size": 512, 00:10:52.395 "num_blocks": 65536, 00:10:52.395 "uuid": "37d0de1d-c2cd-46d3-a5b5-06f0fd623504", 00:10:52.395 "assigned_rate_limits": { 00:10:52.395 "rw_ios_per_sec": 0, 00:10:52.395 "rw_mbytes_per_sec": 0, 00:10:52.395 "r_mbytes_per_sec": 0, 00:10:52.395 "w_mbytes_per_sec": 0 00:10:52.395 }, 00:10:52.395 "claimed": true, 00:10:52.395 "claim_type": "exclusive_write", 00:10:52.395 "zoned": false, 00:10:52.395 "supported_io_types": { 00:10:52.395 "read": true, 00:10:52.395 "write": true, 00:10:52.395 "unmap": true, 00:10:52.395 "flush": true, 00:10:52.395 "reset": true, 00:10:52.395 "nvme_admin": false, 00:10:52.395 "nvme_io": false, 00:10:52.395 "nvme_io_md": false, 00:10:52.395 "write_zeroes": true, 00:10:52.395 "zcopy": true, 00:10:52.395 "get_zone_info": false, 00:10:52.395 "zone_management": false, 00:10:52.395 "zone_append": false, 00:10:52.395 "compare": false, 00:10:52.395 "compare_and_write": false, 00:10:52.395 "abort": true, 00:10:52.395 "seek_hole": false, 00:10:52.395 "seek_data": false, 00:10:52.395 "copy": true, 00:10:52.395 "nvme_iov_md": false 00:10:52.395 }, 00:10:52.395 "memory_domains": [ 00:10:52.395 { 00:10:52.395 "dma_device_id": "system", 00:10:52.395 "dma_device_type": 1 00:10:52.395 }, 00:10:52.395 { 00:10:52.395 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:52.395 "dma_device_type": 2 00:10:52.395 } 00:10:52.395 ], 00:10:52.395 "driver_specific": {} 00:10:52.395 } 00:10:52.395 ] 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:52.395 "name": "Existed_Raid", 00:10:52.395 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.395 "strip_size_kb": 64, 00:10:52.395 "state": "configuring", 00:10:52.395 "raid_level": "raid0", 00:10:52.395 "superblock": false, 00:10:52.395 "num_base_bdevs": 4, 00:10:52.395 "num_base_bdevs_discovered": 1, 00:10:52.395 "num_base_bdevs_operational": 4, 00:10:52.395 "base_bdevs_list": [ 00:10:52.395 { 00:10:52.395 "name": "BaseBdev1", 00:10:52.395 "uuid": "37d0de1d-c2cd-46d3-a5b5-06f0fd623504", 00:10:52.395 "is_configured": true, 00:10:52.395 "data_offset": 0, 00:10:52.395 "data_size": 65536 00:10:52.395 }, 00:10:52.395 { 00:10:52.395 "name": "BaseBdev2", 00:10:52.395 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.395 "is_configured": false, 00:10:52.395 "data_offset": 0, 00:10:52.395 "data_size": 0 00:10:52.395 }, 00:10:52.395 { 00:10:52.395 "name": "BaseBdev3", 00:10:52.395 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.395 "is_configured": false, 00:10:52.395 "data_offset": 0, 00:10:52.395 "data_size": 0 00:10:52.395 }, 00:10:52.395 { 00:10:52.395 "name": "BaseBdev4", 00:10:52.395 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.395 "is_configured": false, 00:10:52.395 "data_offset": 0, 00:10:52.395 "data_size": 0 00:10:52.395 } 00:10:52.395 ] 00:10:52.395 }' 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:52.395 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.652 [2024-11-20 15:58:50.867230] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:52.652 [2024-11-20 15:58:50.867269] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.652 [2024-11-20 15:58:50.875275] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:52.652 [2024-11-20 15:58:50.876868] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:52.652 [2024-11-20 15:58:50.876902] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:52.652 [2024-11-20 15:58:50.876909] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:52.652 [2024-11-20 15:58:50.876917] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:52.652 [2024-11-20 15:58:50.876923] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:52.652 [2024-11-20 15:58:50.876930] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:52.652 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.930 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:52.930 "name": "Existed_Raid", 00:10:52.930 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.930 "strip_size_kb": 64, 00:10:52.930 "state": "configuring", 00:10:52.930 "raid_level": "raid0", 00:10:52.930 "superblock": false, 00:10:52.930 "num_base_bdevs": 4, 00:10:52.930 "num_base_bdevs_discovered": 1, 00:10:52.930 "num_base_bdevs_operational": 4, 00:10:52.930 "base_bdevs_list": [ 00:10:52.930 { 00:10:52.930 "name": "BaseBdev1", 00:10:52.930 "uuid": "37d0de1d-c2cd-46d3-a5b5-06f0fd623504", 00:10:52.930 "is_configured": true, 00:10:52.930 "data_offset": 0, 00:10:52.930 "data_size": 65536 00:10:52.930 }, 00:10:52.930 { 00:10:52.930 "name": "BaseBdev2", 00:10:52.930 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.930 "is_configured": false, 00:10:52.930 "data_offset": 0, 00:10:52.930 "data_size": 0 00:10:52.930 }, 00:10:52.930 { 00:10:52.930 "name": "BaseBdev3", 00:10:52.930 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.930 "is_configured": false, 00:10:52.930 "data_offset": 0, 00:10:52.930 "data_size": 0 00:10:52.930 }, 00:10:52.930 { 00:10:52.930 "name": "BaseBdev4", 00:10:52.930 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.930 "is_configured": false, 00:10:52.930 "data_offset": 0, 00:10:52.930 "data_size": 0 00:10:52.930 } 00:10:52.930 ] 00:10:52.930 }' 00:10:52.930 15:58:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:52.930 15:58:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.200 [2024-11-20 15:58:51.229446] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:53.200 BaseBdev2 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.200 [ 00:10:53.200 { 00:10:53.200 "name": "BaseBdev2", 00:10:53.200 "aliases": [ 00:10:53.200 "d961a4f1-4b43-410e-9c32-27de1f9e5489" 00:10:53.200 ], 00:10:53.200 "product_name": "Malloc disk", 00:10:53.200 "block_size": 512, 00:10:53.200 "num_blocks": 65536, 00:10:53.200 "uuid": "d961a4f1-4b43-410e-9c32-27de1f9e5489", 00:10:53.200 "assigned_rate_limits": { 00:10:53.200 "rw_ios_per_sec": 0, 00:10:53.200 "rw_mbytes_per_sec": 0, 00:10:53.200 "r_mbytes_per_sec": 0, 00:10:53.200 "w_mbytes_per_sec": 0 00:10:53.200 }, 00:10:53.200 "claimed": true, 00:10:53.200 "claim_type": "exclusive_write", 00:10:53.200 "zoned": false, 00:10:53.200 "supported_io_types": { 00:10:53.200 "read": true, 00:10:53.200 "write": true, 00:10:53.200 "unmap": true, 00:10:53.200 "flush": true, 00:10:53.200 "reset": true, 00:10:53.200 "nvme_admin": false, 00:10:53.200 "nvme_io": false, 00:10:53.200 "nvme_io_md": false, 00:10:53.200 "write_zeroes": true, 00:10:53.200 "zcopy": true, 00:10:53.200 "get_zone_info": false, 00:10:53.200 "zone_management": false, 00:10:53.200 "zone_append": false, 00:10:53.200 "compare": false, 00:10:53.200 "compare_and_write": false, 00:10:53.200 "abort": true, 00:10:53.200 "seek_hole": false, 00:10:53.200 "seek_data": false, 00:10:53.200 "copy": true, 00:10:53.200 "nvme_iov_md": false 00:10:53.200 }, 00:10:53.200 "memory_domains": [ 00:10:53.200 { 00:10:53.200 "dma_device_id": "system", 00:10:53.200 "dma_device_type": 1 00:10:53.200 }, 00:10:53.200 { 00:10:53.200 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:53.200 "dma_device_type": 2 00:10:53.200 } 00:10:53.200 ], 00:10:53.200 "driver_specific": {} 00:10:53.200 } 00:10:53.200 ] 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:53.200 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:53.201 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.201 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.201 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.201 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:53.201 "name": "Existed_Raid", 00:10:53.201 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:53.201 "strip_size_kb": 64, 00:10:53.201 "state": "configuring", 00:10:53.201 "raid_level": "raid0", 00:10:53.201 "superblock": false, 00:10:53.201 "num_base_bdevs": 4, 00:10:53.201 "num_base_bdevs_discovered": 2, 00:10:53.201 "num_base_bdevs_operational": 4, 00:10:53.201 "base_bdevs_list": [ 00:10:53.201 { 00:10:53.201 "name": "BaseBdev1", 00:10:53.201 "uuid": "37d0de1d-c2cd-46d3-a5b5-06f0fd623504", 00:10:53.201 "is_configured": true, 00:10:53.201 "data_offset": 0, 00:10:53.201 "data_size": 65536 00:10:53.201 }, 00:10:53.201 { 00:10:53.201 "name": "BaseBdev2", 00:10:53.201 "uuid": "d961a4f1-4b43-410e-9c32-27de1f9e5489", 00:10:53.201 "is_configured": true, 00:10:53.201 "data_offset": 0, 00:10:53.201 "data_size": 65536 00:10:53.201 }, 00:10:53.201 { 00:10:53.201 "name": "BaseBdev3", 00:10:53.201 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:53.201 "is_configured": false, 00:10:53.201 "data_offset": 0, 00:10:53.201 "data_size": 0 00:10:53.201 }, 00:10:53.201 { 00:10:53.201 "name": "BaseBdev4", 00:10:53.201 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:53.201 "is_configured": false, 00:10:53.201 "data_offset": 0, 00:10:53.201 "data_size": 0 00:10:53.201 } 00:10:53.201 ] 00:10:53.201 }' 00:10:53.201 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:53.201 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.461 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:53.461 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.461 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.461 [2024-11-20 15:58:51.602409] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:53.461 BaseBdev3 00:10:53.461 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.461 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:53.461 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:53.461 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:53.461 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:53.461 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:53.461 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:53.461 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:53.461 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.461 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.461 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.461 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:53.461 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.461 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.461 [ 00:10:53.461 { 00:10:53.461 "name": "BaseBdev3", 00:10:53.461 "aliases": [ 00:10:53.461 "7aff4416-1c93-4a9c-a2a1-2d046097e686" 00:10:53.461 ], 00:10:53.461 "product_name": "Malloc disk", 00:10:53.461 "block_size": 512, 00:10:53.461 "num_blocks": 65536, 00:10:53.461 "uuid": "7aff4416-1c93-4a9c-a2a1-2d046097e686", 00:10:53.461 "assigned_rate_limits": { 00:10:53.461 "rw_ios_per_sec": 0, 00:10:53.461 "rw_mbytes_per_sec": 0, 00:10:53.461 "r_mbytes_per_sec": 0, 00:10:53.461 "w_mbytes_per_sec": 0 00:10:53.461 }, 00:10:53.461 "claimed": true, 00:10:53.461 "claim_type": "exclusive_write", 00:10:53.461 "zoned": false, 00:10:53.461 "supported_io_types": { 00:10:53.461 "read": true, 00:10:53.461 "write": true, 00:10:53.461 "unmap": true, 00:10:53.461 "flush": true, 00:10:53.461 "reset": true, 00:10:53.461 "nvme_admin": false, 00:10:53.461 "nvme_io": false, 00:10:53.461 "nvme_io_md": false, 00:10:53.461 "write_zeroes": true, 00:10:53.461 "zcopy": true, 00:10:53.461 "get_zone_info": false, 00:10:53.461 "zone_management": false, 00:10:53.461 "zone_append": false, 00:10:53.461 "compare": false, 00:10:53.461 "compare_and_write": false, 00:10:53.461 "abort": true, 00:10:53.461 "seek_hole": false, 00:10:53.461 "seek_data": false, 00:10:53.461 "copy": true, 00:10:53.461 "nvme_iov_md": false 00:10:53.461 }, 00:10:53.461 "memory_domains": [ 00:10:53.461 { 00:10:53.461 "dma_device_id": "system", 00:10:53.461 "dma_device_type": 1 00:10:53.461 }, 00:10:53.461 { 00:10:53.461 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:53.462 "dma_device_type": 2 00:10:53.462 } 00:10:53.462 ], 00:10:53.462 "driver_specific": {} 00:10:53.462 } 00:10:53.462 ] 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:53.462 "name": "Existed_Raid", 00:10:53.462 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:53.462 "strip_size_kb": 64, 00:10:53.462 "state": "configuring", 00:10:53.462 "raid_level": "raid0", 00:10:53.462 "superblock": false, 00:10:53.462 "num_base_bdevs": 4, 00:10:53.462 "num_base_bdevs_discovered": 3, 00:10:53.462 "num_base_bdevs_operational": 4, 00:10:53.462 "base_bdevs_list": [ 00:10:53.462 { 00:10:53.462 "name": "BaseBdev1", 00:10:53.462 "uuid": "37d0de1d-c2cd-46d3-a5b5-06f0fd623504", 00:10:53.462 "is_configured": true, 00:10:53.462 "data_offset": 0, 00:10:53.462 "data_size": 65536 00:10:53.462 }, 00:10:53.462 { 00:10:53.462 "name": "BaseBdev2", 00:10:53.462 "uuid": "d961a4f1-4b43-410e-9c32-27de1f9e5489", 00:10:53.462 "is_configured": true, 00:10:53.462 "data_offset": 0, 00:10:53.462 "data_size": 65536 00:10:53.462 }, 00:10:53.462 { 00:10:53.462 "name": "BaseBdev3", 00:10:53.462 "uuid": "7aff4416-1c93-4a9c-a2a1-2d046097e686", 00:10:53.462 "is_configured": true, 00:10:53.462 "data_offset": 0, 00:10:53.462 "data_size": 65536 00:10:53.462 }, 00:10:53.462 { 00:10:53.462 "name": "BaseBdev4", 00:10:53.462 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:53.462 "is_configured": false, 00:10:53.462 "data_offset": 0, 00:10:53.462 "data_size": 0 00:10:53.462 } 00:10:53.462 ] 00:10:53.462 }' 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:53.462 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.719 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:53.719 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.719 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.977 [2024-11-20 15:58:51.972973] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:53.977 [2024-11-20 15:58:51.973131] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:10:53.977 [2024-11-20 15:58:51.973158] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:10:53.977 [2024-11-20 15:58:51.973430] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:10:53.977 [2024-11-20 15:58:51.973618] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:10:53.977 [2024-11-20 15:58:51.973710] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:10:53.977 [2024-11-20 15:58:51.973958] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:53.977 BaseBdev4 00:10:53.977 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.977 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:53.977 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:53.977 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:53.977 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:53.977 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:53.977 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:53.977 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:53.977 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.977 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.977 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.977 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:53.977 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.977 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.977 [ 00:10:53.977 { 00:10:53.977 "name": "BaseBdev4", 00:10:53.977 "aliases": [ 00:10:53.977 "23db6200-11d1-4da3-960b-79ff77e2b31a" 00:10:53.977 ], 00:10:53.977 "product_name": "Malloc disk", 00:10:53.977 "block_size": 512, 00:10:53.977 "num_blocks": 65536, 00:10:53.977 "uuid": "23db6200-11d1-4da3-960b-79ff77e2b31a", 00:10:53.978 "assigned_rate_limits": { 00:10:53.978 "rw_ios_per_sec": 0, 00:10:53.978 "rw_mbytes_per_sec": 0, 00:10:53.978 "r_mbytes_per_sec": 0, 00:10:53.978 "w_mbytes_per_sec": 0 00:10:53.978 }, 00:10:53.978 "claimed": true, 00:10:53.978 "claim_type": "exclusive_write", 00:10:53.978 "zoned": false, 00:10:53.978 "supported_io_types": { 00:10:53.978 "read": true, 00:10:53.978 "write": true, 00:10:53.978 "unmap": true, 00:10:53.978 "flush": true, 00:10:53.978 "reset": true, 00:10:53.978 "nvme_admin": false, 00:10:53.978 "nvme_io": false, 00:10:53.978 "nvme_io_md": false, 00:10:53.978 "write_zeroes": true, 00:10:53.978 "zcopy": true, 00:10:53.978 "get_zone_info": false, 00:10:53.978 "zone_management": false, 00:10:53.978 "zone_append": false, 00:10:53.978 "compare": false, 00:10:53.978 "compare_and_write": false, 00:10:53.978 "abort": true, 00:10:53.978 "seek_hole": false, 00:10:53.978 "seek_data": false, 00:10:53.978 "copy": true, 00:10:53.978 "nvme_iov_md": false 00:10:53.978 }, 00:10:53.978 "memory_domains": [ 00:10:53.978 { 00:10:53.978 "dma_device_id": "system", 00:10:53.978 "dma_device_type": 1 00:10:53.978 }, 00:10:53.978 { 00:10:53.978 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:53.978 "dma_device_type": 2 00:10:53.978 } 00:10:53.978 ], 00:10:53.978 "driver_specific": {} 00:10:53.978 } 00:10:53.978 ] 00:10:53.978 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.978 15:58:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:53.978 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:53.978 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:53.978 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:10:53.978 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:53.978 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:53.978 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:53.978 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:53.978 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:53.978 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:53.978 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:53.978 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:53.978 15:58:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:53.978 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:53.978 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:53.978 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.978 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.978 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.978 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:53.978 "name": "Existed_Raid", 00:10:53.978 "uuid": "38bb1708-d1d7-40fb-9239-8d99568082b7", 00:10:53.978 "strip_size_kb": 64, 00:10:53.978 "state": "online", 00:10:53.978 "raid_level": "raid0", 00:10:53.978 "superblock": false, 00:10:53.978 "num_base_bdevs": 4, 00:10:53.978 "num_base_bdevs_discovered": 4, 00:10:53.978 "num_base_bdevs_operational": 4, 00:10:53.978 "base_bdevs_list": [ 00:10:53.978 { 00:10:53.978 "name": "BaseBdev1", 00:10:53.978 "uuid": "37d0de1d-c2cd-46d3-a5b5-06f0fd623504", 00:10:53.978 "is_configured": true, 00:10:53.978 "data_offset": 0, 00:10:53.978 "data_size": 65536 00:10:53.978 }, 00:10:53.978 { 00:10:53.978 "name": "BaseBdev2", 00:10:53.978 "uuid": "d961a4f1-4b43-410e-9c32-27de1f9e5489", 00:10:53.978 "is_configured": true, 00:10:53.978 "data_offset": 0, 00:10:53.978 "data_size": 65536 00:10:53.978 }, 00:10:53.978 { 00:10:53.978 "name": "BaseBdev3", 00:10:53.978 "uuid": "7aff4416-1c93-4a9c-a2a1-2d046097e686", 00:10:53.978 "is_configured": true, 00:10:53.978 "data_offset": 0, 00:10:53.978 "data_size": 65536 00:10:53.978 }, 00:10:53.978 { 00:10:53.978 "name": "BaseBdev4", 00:10:53.978 "uuid": "23db6200-11d1-4da3-960b-79ff77e2b31a", 00:10:53.978 "is_configured": true, 00:10:53.978 "data_offset": 0, 00:10:53.978 "data_size": 65536 00:10:53.978 } 00:10:53.978 ] 00:10:53.978 }' 00:10:53.978 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:53.978 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.236 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:54.236 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:54.236 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:54.236 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:54.236 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:54.236 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:54.236 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:54.236 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.236 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.236 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:54.236 [2024-11-20 15:58:52.321387] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:54.236 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.236 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:54.236 "name": "Existed_Raid", 00:10:54.236 "aliases": [ 00:10:54.236 "38bb1708-d1d7-40fb-9239-8d99568082b7" 00:10:54.236 ], 00:10:54.236 "product_name": "Raid Volume", 00:10:54.236 "block_size": 512, 00:10:54.236 "num_blocks": 262144, 00:10:54.236 "uuid": "38bb1708-d1d7-40fb-9239-8d99568082b7", 00:10:54.236 "assigned_rate_limits": { 00:10:54.236 "rw_ios_per_sec": 0, 00:10:54.236 "rw_mbytes_per_sec": 0, 00:10:54.236 "r_mbytes_per_sec": 0, 00:10:54.236 "w_mbytes_per_sec": 0 00:10:54.236 }, 00:10:54.236 "claimed": false, 00:10:54.236 "zoned": false, 00:10:54.236 "supported_io_types": { 00:10:54.236 "read": true, 00:10:54.236 "write": true, 00:10:54.236 "unmap": true, 00:10:54.236 "flush": true, 00:10:54.236 "reset": true, 00:10:54.236 "nvme_admin": false, 00:10:54.236 "nvme_io": false, 00:10:54.236 "nvme_io_md": false, 00:10:54.236 "write_zeroes": true, 00:10:54.236 "zcopy": false, 00:10:54.236 "get_zone_info": false, 00:10:54.236 "zone_management": false, 00:10:54.236 "zone_append": false, 00:10:54.236 "compare": false, 00:10:54.236 "compare_and_write": false, 00:10:54.236 "abort": false, 00:10:54.236 "seek_hole": false, 00:10:54.236 "seek_data": false, 00:10:54.236 "copy": false, 00:10:54.236 "nvme_iov_md": false 00:10:54.236 }, 00:10:54.236 "memory_domains": [ 00:10:54.236 { 00:10:54.236 "dma_device_id": "system", 00:10:54.236 "dma_device_type": 1 00:10:54.236 }, 00:10:54.236 { 00:10:54.236 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:54.236 "dma_device_type": 2 00:10:54.236 }, 00:10:54.236 { 00:10:54.236 "dma_device_id": "system", 00:10:54.236 "dma_device_type": 1 00:10:54.236 }, 00:10:54.236 { 00:10:54.237 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:54.237 "dma_device_type": 2 00:10:54.237 }, 00:10:54.237 { 00:10:54.237 "dma_device_id": "system", 00:10:54.237 "dma_device_type": 1 00:10:54.237 }, 00:10:54.237 { 00:10:54.237 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:54.237 "dma_device_type": 2 00:10:54.237 }, 00:10:54.237 { 00:10:54.237 "dma_device_id": "system", 00:10:54.237 "dma_device_type": 1 00:10:54.237 }, 00:10:54.237 { 00:10:54.237 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:54.237 "dma_device_type": 2 00:10:54.237 } 00:10:54.237 ], 00:10:54.237 "driver_specific": { 00:10:54.237 "raid": { 00:10:54.237 "uuid": "38bb1708-d1d7-40fb-9239-8d99568082b7", 00:10:54.237 "strip_size_kb": 64, 00:10:54.237 "state": "online", 00:10:54.237 "raid_level": "raid0", 00:10:54.237 "superblock": false, 00:10:54.237 "num_base_bdevs": 4, 00:10:54.237 "num_base_bdevs_discovered": 4, 00:10:54.237 "num_base_bdevs_operational": 4, 00:10:54.237 "base_bdevs_list": [ 00:10:54.237 { 00:10:54.237 "name": "BaseBdev1", 00:10:54.237 "uuid": "37d0de1d-c2cd-46d3-a5b5-06f0fd623504", 00:10:54.237 "is_configured": true, 00:10:54.237 "data_offset": 0, 00:10:54.237 "data_size": 65536 00:10:54.237 }, 00:10:54.237 { 00:10:54.237 "name": "BaseBdev2", 00:10:54.237 "uuid": "d961a4f1-4b43-410e-9c32-27de1f9e5489", 00:10:54.237 "is_configured": true, 00:10:54.237 "data_offset": 0, 00:10:54.237 "data_size": 65536 00:10:54.237 }, 00:10:54.237 { 00:10:54.237 "name": "BaseBdev3", 00:10:54.237 "uuid": "7aff4416-1c93-4a9c-a2a1-2d046097e686", 00:10:54.237 "is_configured": true, 00:10:54.237 "data_offset": 0, 00:10:54.237 "data_size": 65536 00:10:54.237 }, 00:10:54.237 { 00:10:54.237 "name": "BaseBdev4", 00:10:54.237 "uuid": "23db6200-11d1-4da3-960b-79ff77e2b31a", 00:10:54.237 "is_configured": true, 00:10:54.237 "data_offset": 0, 00:10:54.237 "data_size": 65536 00:10:54.237 } 00:10:54.237 ] 00:10:54.237 } 00:10:54.237 } 00:10:54.237 }' 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:54.237 BaseBdev2 00:10:54.237 BaseBdev3 00:10:54.237 BaseBdev4' 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.237 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.495 [2024-11-20 15:58:52.545153] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:54.495 [2024-11-20 15:58:52.545255] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:54.495 [2024-11-20 15:58:52.545340] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:54.495 "name": "Existed_Raid", 00:10:54.495 "uuid": "38bb1708-d1d7-40fb-9239-8d99568082b7", 00:10:54.495 "strip_size_kb": 64, 00:10:54.495 "state": "offline", 00:10:54.495 "raid_level": "raid0", 00:10:54.495 "superblock": false, 00:10:54.495 "num_base_bdevs": 4, 00:10:54.495 "num_base_bdevs_discovered": 3, 00:10:54.495 "num_base_bdevs_operational": 3, 00:10:54.495 "base_bdevs_list": [ 00:10:54.495 { 00:10:54.495 "name": null, 00:10:54.495 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:54.495 "is_configured": false, 00:10:54.495 "data_offset": 0, 00:10:54.495 "data_size": 65536 00:10:54.495 }, 00:10:54.495 { 00:10:54.495 "name": "BaseBdev2", 00:10:54.495 "uuid": "d961a4f1-4b43-410e-9c32-27de1f9e5489", 00:10:54.495 "is_configured": true, 00:10:54.495 "data_offset": 0, 00:10:54.495 "data_size": 65536 00:10:54.495 }, 00:10:54.495 { 00:10:54.495 "name": "BaseBdev3", 00:10:54.495 "uuid": "7aff4416-1c93-4a9c-a2a1-2d046097e686", 00:10:54.495 "is_configured": true, 00:10:54.495 "data_offset": 0, 00:10:54.495 "data_size": 65536 00:10:54.495 }, 00:10:54.495 { 00:10:54.495 "name": "BaseBdev4", 00:10:54.495 "uuid": "23db6200-11d1-4da3-960b-79ff77e2b31a", 00:10:54.495 "is_configured": true, 00:10:54.495 "data_offset": 0, 00:10:54.495 "data_size": 65536 00:10:54.495 } 00:10:54.495 ] 00:10:54.495 }' 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:54.495 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.753 [2024-11-20 15:58:52.943373] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.753 15:58:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.010 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.010 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:55.010 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:55.010 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:55.010 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.010 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.010 [2024-11-20 15:58:53.034065] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:55.010 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.010 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:55.010 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:55.010 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.010 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:55.010 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.010 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.010 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.011 [2024-11-20 15:58:53.121449] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:55.011 [2024-11-20 15:58:53.121555] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.011 BaseBdev2 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.011 [ 00:10:55.011 { 00:10:55.011 "name": "BaseBdev2", 00:10:55.011 "aliases": [ 00:10:55.011 "eed2cb89-6531-4064-b7e8-72ba12050aa7" 00:10:55.011 ], 00:10:55.011 "product_name": "Malloc disk", 00:10:55.011 "block_size": 512, 00:10:55.011 "num_blocks": 65536, 00:10:55.011 "uuid": "eed2cb89-6531-4064-b7e8-72ba12050aa7", 00:10:55.011 "assigned_rate_limits": { 00:10:55.011 "rw_ios_per_sec": 0, 00:10:55.011 "rw_mbytes_per_sec": 0, 00:10:55.011 "r_mbytes_per_sec": 0, 00:10:55.011 "w_mbytes_per_sec": 0 00:10:55.011 }, 00:10:55.011 "claimed": false, 00:10:55.011 "zoned": false, 00:10:55.011 "supported_io_types": { 00:10:55.011 "read": true, 00:10:55.011 "write": true, 00:10:55.011 "unmap": true, 00:10:55.011 "flush": true, 00:10:55.011 "reset": true, 00:10:55.011 "nvme_admin": false, 00:10:55.011 "nvme_io": false, 00:10:55.011 "nvme_io_md": false, 00:10:55.011 "write_zeroes": true, 00:10:55.011 "zcopy": true, 00:10:55.011 "get_zone_info": false, 00:10:55.011 "zone_management": false, 00:10:55.011 "zone_append": false, 00:10:55.011 "compare": false, 00:10:55.011 "compare_and_write": false, 00:10:55.011 "abort": true, 00:10:55.011 "seek_hole": false, 00:10:55.011 "seek_data": false, 00:10:55.011 "copy": true, 00:10:55.011 "nvme_iov_md": false 00:10:55.011 }, 00:10:55.011 "memory_domains": [ 00:10:55.011 { 00:10:55.011 "dma_device_id": "system", 00:10:55.011 "dma_device_type": 1 00:10:55.011 }, 00:10:55.011 { 00:10:55.011 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:55.011 "dma_device_type": 2 00:10:55.011 } 00:10:55.011 ], 00:10:55.011 "driver_specific": {} 00:10:55.011 } 00:10:55.011 ] 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.011 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.269 BaseBdev3 00:10:55.269 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.269 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:55.269 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:55.269 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:55.269 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:55.269 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:55.269 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:55.269 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:55.269 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.269 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.269 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.269 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:55.269 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.270 [ 00:10:55.270 { 00:10:55.270 "name": "BaseBdev3", 00:10:55.270 "aliases": [ 00:10:55.270 "5fc2bd8a-a94f-4f50-a0e7-0d13cc794485" 00:10:55.270 ], 00:10:55.270 "product_name": "Malloc disk", 00:10:55.270 "block_size": 512, 00:10:55.270 "num_blocks": 65536, 00:10:55.270 "uuid": "5fc2bd8a-a94f-4f50-a0e7-0d13cc794485", 00:10:55.270 "assigned_rate_limits": { 00:10:55.270 "rw_ios_per_sec": 0, 00:10:55.270 "rw_mbytes_per_sec": 0, 00:10:55.270 "r_mbytes_per_sec": 0, 00:10:55.270 "w_mbytes_per_sec": 0 00:10:55.270 }, 00:10:55.270 "claimed": false, 00:10:55.270 "zoned": false, 00:10:55.270 "supported_io_types": { 00:10:55.270 "read": true, 00:10:55.270 "write": true, 00:10:55.270 "unmap": true, 00:10:55.270 "flush": true, 00:10:55.270 "reset": true, 00:10:55.270 "nvme_admin": false, 00:10:55.270 "nvme_io": false, 00:10:55.270 "nvme_io_md": false, 00:10:55.270 "write_zeroes": true, 00:10:55.270 "zcopy": true, 00:10:55.270 "get_zone_info": false, 00:10:55.270 "zone_management": false, 00:10:55.270 "zone_append": false, 00:10:55.270 "compare": false, 00:10:55.270 "compare_and_write": false, 00:10:55.270 "abort": true, 00:10:55.270 "seek_hole": false, 00:10:55.270 "seek_data": false, 00:10:55.270 "copy": true, 00:10:55.270 "nvme_iov_md": false 00:10:55.270 }, 00:10:55.270 "memory_domains": [ 00:10:55.270 { 00:10:55.270 "dma_device_id": "system", 00:10:55.270 "dma_device_type": 1 00:10:55.270 }, 00:10:55.270 { 00:10:55.270 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:55.270 "dma_device_type": 2 00:10:55.270 } 00:10:55.270 ], 00:10:55.270 "driver_specific": {} 00:10:55.270 } 00:10:55.270 ] 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.270 BaseBdev4 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.270 [ 00:10:55.270 { 00:10:55.270 "name": "BaseBdev4", 00:10:55.270 "aliases": [ 00:10:55.270 "a2500ff1-6276-49b2-a3f2-31f6406cd8f1" 00:10:55.270 ], 00:10:55.270 "product_name": "Malloc disk", 00:10:55.270 "block_size": 512, 00:10:55.270 "num_blocks": 65536, 00:10:55.270 "uuid": "a2500ff1-6276-49b2-a3f2-31f6406cd8f1", 00:10:55.270 "assigned_rate_limits": { 00:10:55.270 "rw_ios_per_sec": 0, 00:10:55.270 "rw_mbytes_per_sec": 0, 00:10:55.270 "r_mbytes_per_sec": 0, 00:10:55.270 "w_mbytes_per_sec": 0 00:10:55.270 }, 00:10:55.270 "claimed": false, 00:10:55.270 "zoned": false, 00:10:55.270 "supported_io_types": { 00:10:55.270 "read": true, 00:10:55.270 "write": true, 00:10:55.270 "unmap": true, 00:10:55.270 "flush": true, 00:10:55.270 "reset": true, 00:10:55.270 "nvme_admin": false, 00:10:55.270 "nvme_io": false, 00:10:55.270 "nvme_io_md": false, 00:10:55.270 "write_zeroes": true, 00:10:55.270 "zcopy": true, 00:10:55.270 "get_zone_info": false, 00:10:55.270 "zone_management": false, 00:10:55.270 "zone_append": false, 00:10:55.270 "compare": false, 00:10:55.270 "compare_and_write": false, 00:10:55.270 "abort": true, 00:10:55.270 "seek_hole": false, 00:10:55.270 "seek_data": false, 00:10:55.270 "copy": true, 00:10:55.270 "nvme_iov_md": false 00:10:55.270 }, 00:10:55.270 "memory_domains": [ 00:10:55.270 { 00:10:55.270 "dma_device_id": "system", 00:10:55.270 "dma_device_type": 1 00:10:55.270 }, 00:10:55.270 { 00:10:55.270 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:55.270 "dma_device_type": 2 00:10:55.270 } 00:10:55.270 ], 00:10:55.270 "driver_specific": {} 00:10:55.270 } 00:10:55.270 ] 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.270 [2024-11-20 15:58:53.355129] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:55.270 [2024-11-20 15:58:53.355246] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:55.270 [2024-11-20 15:58:53.355307] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:55.270 [2024-11-20 15:58:53.356886] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:55.270 [2024-11-20 15:58:53.356994] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:55.270 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:55.271 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:55.271 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:55.271 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:55.271 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:55.271 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:55.271 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.271 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:55.271 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.271 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.271 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.271 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:55.271 "name": "Existed_Raid", 00:10:55.271 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:55.271 "strip_size_kb": 64, 00:10:55.271 "state": "configuring", 00:10:55.271 "raid_level": "raid0", 00:10:55.271 "superblock": false, 00:10:55.271 "num_base_bdevs": 4, 00:10:55.271 "num_base_bdevs_discovered": 3, 00:10:55.271 "num_base_bdevs_operational": 4, 00:10:55.271 "base_bdevs_list": [ 00:10:55.271 { 00:10:55.271 "name": "BaseBdev1", 00:10:55.271 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:55.271 "is_configured": false, 00:10:55.271 "data_offset": 0, 00:10:55.271 "data_size": 0 00:10:55.271 }, 00:10:55.271 { 00:10:55.271 "name": "BaseBdev2", 00:10:55.271 "uuid": "eed2cb89-6531-4064-b7e8-72ba12050aa7", 00:10:55.271 "is_configured": true, 00:10:55.271 "data_offset": 0, 00:10:55.271 "data_size": 65536 00:10:55.271 }, 00:10:55.271 { 00:10:55.271 "name": "BaseBdev3", 00:10:55.271 "uuid": "5fc2bd8a-a94f-4f50-a0e7-0d13cc794485", 00:10:55.271 "is_configured": true, 00:10:55.271 "data_offset": 0, 00:10:55.271 "data_size": 65536 00:10:55.271 }, 00:10:55.271 { 00:10:55.271 "name": "BaseBdev4", 00:10:55.271 "uuid": "a2500ff1-6276-49b2-a3f2-31f6406cd8f1", 00:10:55.271 "is_configured": true, 00:10:55.271 "data_offset": 0, 00:10:55.271 "data_size": 65536 00:10:55.271 } 00:10:55.271 ] 00:10:55.271 }' 00:10:55.271 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:55.271 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.529 [2024-11-20 15:58:53.675234] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:55.529 "name": "Existed_Raid", 00:10:55.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:55.529 "strip_size_kb": 64, 00:10:55.529 "state": "configuring", 00:10:55.529 "raid_level": "raid0", 00:10:55.529 "superblock": false, 00:10:55.529 "num_base_bdevs": 4, 00:10:55.529 "num_base_bdevs_discovered": 2, 00:10:55.529 "num_base_bdevs_operational": 4, 00:10:55.529 "base_bdevs_list": [ 00:10:55.529 { 00:10:55.529 "name": "BaseBdev1", 00:10:55.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:55.529 "is_configured": false, 00:10:55.529 "data_offset": 0, 00:10:55.529 "data_size": 0 00:10:55.529 }, 00:10:55.529 { 00:10:55.529 "name": null, 00:10:55.529 "uuid": "eed2cb89-6531-4064-b7e8-72ba12050aa7", 00:10:55.529 "is_configured": false, 00:10:55.529 "data_offset": 0, 00:10:55.529 "data_size": 65536 00:10:55.529 }, 00:10:55.529 { 00:10:55.529 "name": "BaseBdev3", 00:10:55.529 "uuid": "5fc2bd8a-a94f-4f50-a0e7-0d13cc794485", 00:10:55.529 "is_configured": true, 00:10:55.529 "data_offset": 0, 00:10:55.529 "data_size": 65536 00:10:55.529 }, 00:10:55.529 { 00:10:55.529 "name": "BaseBdev4", 00:10:55.529 "uuid": "a2500ff1-6276-49b2-a3f2-31f6406cd8f1", 00:10:55.529 "is_configured": true, 00:10:55.529 "data_offset": 0, 00:10:55.529 "data_size": 65536 00:10:55.529 } 00:10:55.529 ] 00:10:55.529 }' 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:55.529 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.787 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.787 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:55.787 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.787 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.787 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.787 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:55.787 15:58:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:55.787 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.787 15:58:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.787 [2024-11-20 15:58:54.017423] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:55.787 BaseBdev1 00:10:55.787 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.787 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:55.787 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:55.787 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:55.787 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:55.787 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:55.787 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:55.787 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:55.787 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.787 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.787 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.787 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:55.787 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.787 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.045 [ 00:10:56.045 { 00:10:56.045 "name": "BaseBdev1", 00:10:56.045 "aliases": [ 00:10:56.045 "42c10ef8-d7bc-4438-88d1-b9deb138c7a3" 00:10:56.045 ], 00:10:56.045 "product_name": "Malloc disk", 00:10:56.045 "block_size": 512, 00:10:56.045 "num_blocks": 65536, 00:10:56.045 "uuid": "42c10ef8-d7bc-4438-88d1-b9deb138c7a3", 00:10:56.045 "assigned_rate_limits": { 00:10:56.045 "rw_ios_per_sec": 0, 00:10:56.045 "rw_mbytes_per_sec": 0, 00:10:56.045 "r_mbytes_per_sec": 0, 00:10:56.045 "w_mbytes_per_sec": 0 00:10:56.045 }, 00:10:56.045 "claimed": true, 00:10:56.045 "claim_type": "exclusive_write", 00:10:56.045 "zoned": false, 00:10:56.045 "supported_io_types": { 00:10:56.045 "read": true, 00:10:56.045 "write": true, 00:10:56.045 "unmap": true, 00:10:56.045 "flush": true, 00:10:56.045 "reset": true, 00:10:56.045 "nvme_admin": false, 00:10:56.045 "nvme_io": false, 00:10:56.045 "nvme_io_md": false, 00:10:56.045 "write_zeroes": true, 00:10:56.045 "zcopy": true, 00:10:56.045 "get_zone_info": false, 00:10:56.045 "zone_management": false, 00:10:56.045 "zone_append": false, 00:10:56.045 "compare": false, 00:10:56.045 "compare_and_write": false, 00:10:56.045 "abort": true, 00:10:56.045 "seek_hole": false, 00:10:56.045 "seek_data": false, 00:10:56.045 "copy": true, 00:10:56.045 "nvme_iov_md": false 00:10:56.045 }, 00:10:56.045 "memory_domains": [ 00:10:56.045 { 00:10:56.045 "dma_device_id": "system", 00:10:56.045 "dma_device_type": 1 00:10:56.045 }, 00:10:56.045 { 00:10:56.045 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:56.045 "dma_device_type": 2 00:10:56.045 } 00:10:56.045 ], 00:10:56.045 "driver_specific": {} 00:10:56.045 } 00:10:56.045 ] 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:56.045 "name": "Existed_Raid", 00:10:56.045 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:56.045 "strip_size_kb": 64, 00:10:56.045 "state": "configuring", 00:10:56.045 "raid_level": "raid0", 00:10:56.045 "superblock": false, 00:10:56.045 "num_base_bdevs": 4, 00:10:56.045 "num_base_bdevs_discovered": 3, 00:10:56.045 "num_base_bdevs_operational": 4, 00:10:56.045 "base_bdevs_list": [ 00:10:56.045 { 00:10:56.045 "name": "BaseBdev1", 00:10:56.045 "uuid": "42c10ef8-d7bc-4438-88d1-b9deb138c7a3", 00:10:56.045 "is_configured": true, 00:10:56.045 "data_offset": 0, 00:10:56.045 "data_size": 65536 00:10:56.045 }, 00:10:56.045 { 00:10:56.045 "name": null, 00:10:56.045 "uuid": "eed2cb89-6531-4064-b7e8-72ba12050aa7", 00:10:56.045 "is_configured": false, 00:10:56.045 "data_offset": 0, 00:10:56.045 "data_size": 65536 00:10:56.045 }, 00:10:56.045 { 00:10:56.045 "name": "BaseBdev3", 00:10:56.045 "uuid": "5fc2bd8a-a94f-4f50-a0e7-0d13cc794485", 00:10:56.045 "is_configured": true, 00:10:56.045 "data_offset": 0, 00:10:56.045 "data_size": 65536 00:10:56.045 }, 00:10:56.045 { 00:10:56.045 "name": "BaseBdev4", 00:10:56.045 "uuid": "a2500ff1-6276-49b2-a3f2-31f6406cd8f1", 00:10:56.045 "is_configured": true, 00:10:56.045 "data_offset": 0, 00:10:56.045 "data_size": 65536 00:10:56.045 } 00:10:56.045 ] 00:10:56.045 }' 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:56.045 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.324 [2024-11-20 15:58:54.385569] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:56.324 "name": "Existed_Raid", 00:10:56.324 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:56.324 "strip_size_kb": 64, 00:10:56.324 "state": "configuring", 00:10:56.324 "raid_level": "raid0", 00:10:56.324 "superblock": false, 00:10:56.324 "num_base_bdevs": 4, 00:10:56.324 "num_base_bdevs_discovered": 2, 00:10:56.324 "num_base_bdevs_operational": 4, 00:10:56.324 "base_bdevs_list": [ 00:10:56.324 { 00:10:56.324 "name": "BaseBdev1", 00:10:56.324 "uuid": "42c10ef8-d7bc-4438-88d1-b9deb138c7a3", 00:10:56.324 "is_configured": true, 00:10:56.324 "data_offset": 0, 00:10:56.324 "data_size": 65536 00:10:56.324 }, 00:10:56.324 { 00:10:56.324 "name": null, 00:10:56.324 "uuid": "eed2cb89-6531-4064-b7e8-72ba12050aa7", 00:10:56.324 "is_configured": false, 00:10:56.324 "data_offset": 0, 00:10:56.324 "data_size": 65536 00:10:56.324 }, 00:10:56.324 { 00:10:56.324 "name": null, 00:10:56.324 "uuid": "5fc2bd8a-a94f-4f50-a0e7-0d13cc794485", 00:10:56.324 "is_configured": false, 00:10:56.324 "data_offset": 0, 00:10:56.324 "data_size": 65536 00:10:56.324 }, 00:10:56.324 { 00:10:56.324 "name": "BaseBdev4", 00:10:56.324 "uuid": "a2500ff1-6276-49b2-a3f2-31f6406cd8f1", 00:10:56.324 "is_configured": true, 00:10:56.324 "data_offset": 0, 00:10:56.324 "data_size": 65536 00:10:56.324 } 00:10:56.324 ] 00:10:56.324 }' 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:56.324 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.582 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:56.582 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:56.582 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.582 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.582 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.582 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:56.582 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:56.582 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.583 [2024-11-20 15:58:54.753631] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:56.583 "name": "Existed_Raid", 00:10:56.583 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:56.583 "strip_size_kb": 64, 00:10:56.583 "state": "configuring", 00:10:56.583 "raid_level": "raid0", 00:10:56.583 "superblock": false, 00:10:56.583 "num_base_bdevs": 4, 00:10:56.583 "num_base_bdevs_discovered": 3, 00:10:56.583 "num_base_bdevs_operational": 4, 00:10:56.583 "base_bdevs_list": [ 00:10:56.583 { 00:10:56.583 "name": "BaseBdev1", 00:10:56.583 "uuid": "42c10ef8-d7bc-4438-88d1-b9deb138c7a3", 00:10:56.583 "is_configured": true, 00:10:56.583 "data_offset": 0, 00:10:56.583 "data_size": 65536 00:10:56.583 }, 00:10:56.583 { 00:10:56.583 "name": null, 00:10:56.583 "uuid": "eed2cb89-6531-4064-b7e8-72ba12050aa7", 00:10:56.583 "is_configured": false, 00:10:56.583 "data_offset": 0, 00:10:56.583 "data_size": 65536 00:10:56.583 }, 00:10:56.583 { 00:10:56.583 "name": "BaseBdev3", 00:10:56.583 "uuid": "5fc2bd8a-a94f-4f50-a0e7-0d13cc794485", 00:10:56.583 "is_configured": true, 00:10:56.583 "data_offset": 0, 00:10:56.583 "data_size": 65536 00:10:56.583 }, 00:10:56.583 { 00:10:56.583 "name": "BaseBdev4", 00:10:56.583 "uuid": "a2500ff1-6276-49b2-a3f2-31f6406cd8f1", 00:10:56.583 "is_configured": true, 00:10:56.583 "data_offset": 0, 00:10:56.583 "data_size": 65536 00:10:56.583 } 00:10:56.583 ] 00:10:56.583 }' 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:56.583 15:58:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.153 [2024-11-20 15:58:55.161734] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:57.153 "name": "Existed_Raid", 00:10:57.153 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:57.153 "strip_size_kb": 64, 00:10:57.153 "state": "configuring", 00:10:57.153 "raid_level": "raid0", 00:10:57.153 "superblock": false, 00:10:57.153 "num_base_bdevs": 4, 00:10:57.153 "num_base_bdevs_discovered": 2, 00:10:57.153 "num_base_bdevs_operational": 4, 00:10:57.153 "base_bdevs_list": [ 00:10:57.153 { 00:10:57.153 "name": null, 00:10:57.153 "uuid": "42c10ef8-d7bc-4438-88d1-b9deb138c7a3", 00:10:57.153 "is_configured": false, 00:10:57.153 "data_offset": 0, 00:10:57.153 "data_size": 65536 00:10:57.153 }, 00:10:57.153 { 00:10:57.153 "name": null, 00:10:57.153 "uuid": "eed2cb89-6531-4064-b7e8-72ba12050aa7", 00:10:57.153 "is_configured": false, 00:10:57.153 "data_offset": 0, 00:10:57.153 "data_size": 65536 00:10:57.153 }, 00:10:57.153 { 00:10:57.153 "name": "BaseBdev3", 00:10:57.153 "uuid": "5fc2bd8a-a94f-4f50-a0e7-0d13cc794485", 00:10:57.153 "is_configured": true, 00:10:57.153 "data_offset": 0, 00:10:57.153 "data_size": 65536 00:10:57.153 }, 00:10:57.153 { 00:10:57.153 "name": "BaseBdev4", 00:10:57.153 "uuid": "a2500ff1-6276-49b2-a3f2-31f6406cd8f1", 00:10:57.153 "is_configured": true, 00:10:57.153 "data_offset": 0, 00:10:57.153 "data_size": 65536 00:10:57.153 } 00:10:57.153 ] 00:10:57.153 }' 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:57.153 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.411 [2024-11-20 15:58:55.540052] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:57.411 "name": "Existed_Raid", 00:10:57.411 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:57.411 "strip_size_kb": 64, 00:10:57.411 "state": "configuring", 00:10:57.411 "raid_level": "raid0", 00:10:57.411 "superblock": false, 00:10:57.411 "num_base_bdevs": 4, 00:10:57.411 "num_base_bdevs_discovered": 3, 00:10:57.411 "num_base_bdevs_operational": 4, 00:10:57.411 "base_bdevs_list": [ 00:10:57.411 { 00:10:57.411 "name": null, 00:10:57.411 "uuid": "42c10ef8-d7bc-4438-88d1-b9deb138c7a3", 00:10:57.411 "is_configured": false, 00:10:57.411 "data_offset": 0, 00:10:57.411 "data_size": 65536 00:10:57.411 }, 00:10:57.411 { 00:10:57.411 "name": "BaseBdev2", 00:10:57.411 "uuid": "eed2cb89-6531-4064-b7e8-72ba12050aa7", 00:10:57.411 "is_configured": true, 00:10:57.411 "data_offset": 0, 00:10:57.411 "data_size": 65536 00:10:57.411 }, 00:10:57.411 { 00:10:57.411 "name": "BaseBdev3", 00:10:57.411 "uuid": "5fc2bd8a-a94f-4f50-a0e7-0d13cc794485", 00:10:57.411 "is_configured": true, 00:10:57.411 "data_offset": 0, 00:10:57.411 "data_size": 65536 00:10:57.411 }, 00:10:57.411 { 00:10:57.411 "name": "BaseBdev4", 00:10:57.411 "uuid": "a2500ff1-6276-49b2-a3f2-31f6406cd8f1", 00:10:57.411 "is_configured": true, 00:10:57.411 "data_offset": 0, 00:10:57.411 "data_size": 65536 00:10:57.411 } 00:10:57.411 ] 00:10:57.411 }' 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:57.411 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.669 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:57.669 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.669 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.669 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:57.669 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.669 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:57.669 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:57.669 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:57.669 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.669 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.669 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 42c10ef8-d7bc-4438-88d1-b9deb138c7a3 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.927 [2024-11-20 15:58:55.958707] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:57.927 NewBaseBdev 00:10:57.927 [2024-11-20 15:58:55.958886] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:10:57.927 [2024-11-20 15:58:55.958898] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:10:57.927 [2024-11-20 15:58:55.959121] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:10:57.927 [2024-11-20 15:58:55.959230] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:10:57.927 [2024-11-20 15:58:55.959238] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:10:57.927 [2024-11-20 15:58:55.959424] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.927 [ 00:10:57.927 { 00:10:57.927 "name": "NewBaseBdev", 00:10:57.927 "aliases": [ 00:10:57.927 "42c10ef8-d7bc-4438-88d1-b9deb138c7a3" 00:10:57.927 ], 00:10:57.927 "product_name": "Malloc disk", 00:10:57.927 "block_size": 512, 00:10:57.927 "num_blocks": 65536, 00:10:57.927 "uuid": "42c10ef8-d7bc-4438-88d1-b9deb138c7a3", 00:10:57.927 "assigned_rate_limits": { 00:10:57.927 "rw_ios_per_sec": 0, 00:10:57.927 "rw_mbytes_per_sec": 0, 00:10:57.927 "r_mbytes_per_sec": 0, 00:10:57.927 "w_mbytes_per_sec": 0 00:10:57.927 }, 00:10:57.927 "claimed": true, 00:10:57.927 "claim_type": "exclusive_write", 00:10:57.927 "zoned": false, 00:10:57.927 "supported_io_types": { 00:10:57.927 "read": true, 00:10:57.927 "write": true, 00:10:57.927 "unmap": true, 00:10:57.927 "flush": true, 00:10:57.927 "reset": true, 00:10:57.927 "nvme_admin": false, 00:10:57.927 "nvme_io": false, 00:10:57.927 "nvme_io_md": false, 00:10:57.927 "write_zeroes": true, 00:10:57.927 "zcopy": true, 00:10:57.927 "get_zone_info": false, 00:10:57.927 "zone_management": false, 00:10:57.927 "zone_append": false, 00:10:57.927 "compare": false, 00:10:57.927 "compare_and_write": false, 00:10:57.927 "abort": true, 00:10:57.927 "seek_hole": false, 00:10:57.927 "seek_data": false, 00:10:57.927 "copy": true, 00:10:57.927 "nvme_iov_md": false 00:10:57.927 }, 00:10:57.927 "memory_domains": [ 00:10:57.927 { 00:10:57.927 "dma_device_id": "system", 00:10:57.927 "dma_device_type": 1 00:10:57.927 }, 00:10:57.927 { 00:10:57.927 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:57.927 "dma_device_type": 2 00:10:57.927 } 00:10:57.927 ], 00:10:57.927 "driver_specific": {} 00:10:57.927 } 00:10:57.927 ] 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:57.927 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.928 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.928 15:58:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.928 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:57.928 "name": "Existed_Raid", 00:10:57.928 "uuid": "d50ed302-f061-4a9f-970c-c47e921f0b54", 00:10:57.928 "strip_size_kb": 64, 00:10:57.928 "state": "online", 00:10:57.928 "raid_level": "raid0", 00:10:57.928 "superblock": false, 00:10:57.928 "num_base_bdevs": 4, 00:10:57.928 "num_base_bdevs_discovered": 4, 00:10:57.928 "num_base_bdevs_operational": 4, 00:10:57.928 "base_bdevs_list": [ 00:10:57.928 { 00:10:57.928 "name": "NewBaseBdev", 00:10:57.928 "uuid": "42c10ef8-d7bc-4438-88d1-b9deb138c7a3", 00:10:57.928 "is_configured": true, 00:10:57.928 "data_offset": 0, 00:10:57.928 "data_size": 65536 00:10:57.928 }, 00:10:57.928 { 00:10:57.928 "name": "BaseBdev2", 00:10:57.928 "uuid": "eed2cb89-6531-4064-b7e8-72ba12050aa7", 00:10:57.928 "is_configured": true, 00:10:57.928 "data_offset": 0, 00:10:57.928 "data_size": 65536 00:10:57.928 }, 00:10:57.928 { 00:10:57.928 "name": "BaseBdev3", 00:10:57.928 "uuid": "5fc2bd8a-a94f-4f50-a0e7-0d13cc794485", 00:10:57.928 "is_configured": true, 00:10:57.928 "data_offset": 0, 00:10:57.928 "data_size": 65536 00:10:57.928 }, 00:10:57.928 { 00:10:57.928 "name": "BaseBdev4", 00:10:57.928 "uuid": "a2500ff1-6276-49b2-a3f2-31f6406cd8f1", 00:10:57.928 "is_configured": true, 00:10:57.928 "data_offset": 0, 00:10:57.928 "data_size": 65536 00:10:57.928 } 00:10:57.928 ] 00:10:57.928 }' 00:10:57.928 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:57.928 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:58.186 [2024-11-20 15:58:56.299119] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:58.186 "name": "Existed_Raid", 00:10:58.186 "aliases": [ 00:10:58.186 "d50ed302-f061-4a9f-970c-c47e921f0b54" 00:10:58.186 ], 00:10:58.186 "product_name": "Raid Volume", 00:10:58.186 "block_size": 512, 00:10:58.186 "num_blocks": 262144, 00:10:58.186 "uuid": "d50ed302-f061-4a9f-970c-c47e921f0b54", 00:10:58.186 "assigned_rate_limits": { 00:10:58.186 "rw_ios_per_sec": 0, 00:10:58.186 "rw_mbytes_per_sec": 0, 00:10:58.186 "r_mbytes_per_sec": 0, 00:10:58.186 "w_mbytes_per_sec": 0 00:10:58.186 }, 00:10:58.186 "claimed": false, 00:10:58.186 "zoned": false, 00:10:58.186 "supported_io_types": { 00:10:58.186 "read": true, 00:10:58.186 "write": true, 00:10:58.186 "unmap": true, 00:10:58.186 "flush": true, 00:10:58.186 "reset": true, 00:10:58.186 "nvme_admin": false, 00:10:58.186 "nvme_io": false, 00:10:58.186 "nvme_io_md": false, 00:10:58.186 "write_zeroes": true, 00:10:58.186 "zcopy": false, 00:10:58.186 "get_zone_info": false, 00:10:58.186 "zone_management": false, 00:10:58.186 "zone_append": false, 00:10:58.186 "compare": false, 00:10:58.186 "compare_and_write": false, 00:10:58.186 "abort": false, 00:10:58.186 "seek_hole": false, 00:10:58.186 "seek_data": false, 00:10:58.186 "copy": false, 00:10:58.186 "nvme_iov_md": false 00:10:58.186 }, 00:10:58.186 "memory_domains": [ 00:10:58.186 { 00:10:58.186 "dma_device_id": "system", 00:10:58.186 "dma_device_type": 1 00:10:58.186 }, 00:10:58.186 { 00:10:58.186 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:58.186 "dma_device_type": 2 00:10:58.186 }, 00:10:58.186 { 00:10:58.186 "dma_device_id": "system", 00:10:58.186 "dma_device_type": 1 00:10:58.186 }, 00:10:58.186 { 00:10:58.186 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:58.186 "dma_device_type": 2 00:10:58.186 }, 00:10:58.186 { 00:10:58.186 "dma_device_id": "system", 00:10:58.186 "dma_device_type": 1 00:10:58.186 }, 00:10:58.186 { 00:10:58.186 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:58.186 "dma_device_type": 2 00:10:58.186 }, 00:10:58.186 { 00:10:58.186 "dma_device_id": "system", 00:10:58.186 "dma_device_type": 1 00:10:58.186 }, 00:10:58.186 { 00:10:58.186 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:58.186 "dma_device_type": 2 00:10:58.186 } 00:10:58.186 ], 00:10:58.186 "driver_specific": { 00:10:58.186 "raid": { 00:10:58.186 "uuid": "d50ed302-f061-4a9f-970c-c47e921f0b54", 00:10:58.186 "strip_size_kb": 64, 00:10:58.186 "state": "online", 00:10:58.186 "raid_level": "raid0", 00:10:58.186 "superblock": false, 00:10:58.186 "num_base_bdevs": 4, 00:10:58.186 "num_base_bdevs_discovered": 4, 00:10:58.186 "num_base_bdevs_operational": 4, 00:10:58.186 "base_bdevs_list": [ 00:10:58.186 { 00:10:58.186 "name": "NewBaseBdev", 00:10:58.186 "uuid": "42c10ef8-d7bc-4438-88d1-b9deb138c7a3", 00:10:58.186 "is_configured": true, 00:10:58.186 "data_offset": 0, 00:10:58.186 "data_size": 65536 00:10:58.186 }, 00:10:58.186 { 00:10:58.186 "name": "BaseBdev2", 00:10:58.186 "uuid": "eed2cb89-6531-4064-b7e8-72ba12050aa7", 00:10:58.186 "is_configured": true, 00:10:58.186 "data_offset": 0, 00:10:58.186 "data_size": 65536 00:10:58.186 }, 00:10:58.186 { 00:10:58.186 "name": "BaseBdev3", 00:10:58.186 "uuid": "5fc2bd8a-a94f-4f50-a0e7-0d13cc794485", 00:10:58.186 "is_configured": true, 00:10:58.186 "data_offset": 0, 00:10:58.186 "data_size": 65536 00:10:58.186 }, 00:10:58.186 { 00:10:58.186 "name": "BaseBdev4", 00:10:58.186 "uuid": "a2500ff1-6276-49b2-a3f2-31f6406cd8f1", 00:10:58.186 "is_configured": true, 00:10:58.186 "data_offset": 0, 00:10:58.186 "data_size": 65536 00:10:58.186 } 00:10:58.186 ] 00:10:58.186 } 00:10:58.186 } 00:10:58.186 }' 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:58.186 BaseBdev2 00:10:58.186 BaseBdev3 00:10:58.186 BaseBdev4' 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:58.186 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:58.187 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.187 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.187 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:58.187 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.445 [2024-11-20 15:58:56.514843] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:58.445 [2024-11-20 15:58:56.514942] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:58.445 [2024-11-20 15:58:56.515039] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:58.445 [2024-11-20 15:58:56.515195] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:58.445 [2024-11-20 15:58:56.515300] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 67567 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 67567 ']' 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 67567 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 67567 00:10:58.445 killing process with pid 67567 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 67567' 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 67567 00:10:58.445 [2024-11-20 15:58:56.546011] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:58.445 15:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 67567 00:10:58.702 [2024-11-20 15:58:56.738601] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:59.269 ************************************ 00:10:59.269 END TEST raid_state_function_test 00:10:59.269 ************************************ 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:10:59.269 00:10:59.269 real 0m8.003s 00:10:59.269 user 0m12.958s 00:10:59.269 sys 0m1.309s 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.269 15:58:57 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 4 true 00:10:59.269 15:58:57 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:59.269 15:58:57 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:59.269 15:58:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:59.269 ************************************ 00:10:59.269 START TEST raid_state_function_test_sb 00:10:59.269 ************************************ 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 4 true 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=68201 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 68201' 00:10:59.269 Process raid pid: 68201 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 68201 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 68201 ']' 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:59.269 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:59.269 15:58:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:59.269 [2024-11-20 15:58:57.419392] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:10:59.269 [2024-11-20 15:58:57.419678] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:59.527 [2024-11-20 15:58:57.577594] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:59.527 [2024-11-20 15:58:57.660949] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:59.527 [2024-11-20 15:58:57.772401] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:59.527 [2024-11-20 15:58:57.772536] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.093 [2024-11-20 15:58:58.292047] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:00.093 [2024-11-20 15:58:58.292175] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:00.093 [2024-11-20 15:58:58.292232] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:00.093 [2024-11-20 15:58:58.292254] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:00.093 [2024-11-20 15:58:58.292269] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:00.093 [2024-11-20 15:58:58.292286] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:00.093 [2024-11-20 15:58:58.292331] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:00.093 [2024-11-20 15:58:58.292352] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:00.093 "name": "Existed_Raid", 00:11:00.093 "uuid": "75e9fbc0-41e3-49b8-8fb8-9f78dee417e4", 00:11:00.093 "strip_size_kb": 64, 00:11:00.093 "state": "configuring", 00:11:00.093 "raid_level": "raid0", 00:11:00.093 "superblock": true, 00:11:00.093 "num_base_bdevs": 4, 00:11:00.093 "num_base_bdevs_discovered": 0, 00:11:00.093 "num_base_bdevs_operational": 4, 00:11:00.093 "base_bdevs_list": [ 00:11:00.093 { 00:11:00.093 "name": "BaseBdev1", 00:11:00.093 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.093 "is_configured": false, 00:11:00.093 "data_offset": 0, 00:11:00.093 "data_size": 0 00:11:00.093 }, 00:11:00.093 { 00:11:00.093 "name": "BaseBdev2", 00:11:00.093 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.093 "is_configured": false, 00:11:00.093 "data_offset": 0, 00:11:00.093 "data_size": 0 00:11:00.093 }, 00:11:00.093 { 00:11:00.093 "name": "BaseBdev3", 00:11:00.093 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.093 "is_configured": false, 00:11:00.093 "data_offset": 0, 00:11:00.093 "data_size": 0 00:11:00.093 }, 00:11:00.093 { 00:11:00.093 "name": "BaseBdev4", 00:11:00.093 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.093 "is_configured": false, 00:11:00.093 "data_offset": 0, 00:11:00.093 "data_size": 0 00:11:00.093 } 00:11:00.093 ] 00:11:00.093 }' 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:00.093 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.351 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:00.351 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.351 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.609 [2024-11-20 15:58:58.604054] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:00.609 [2024-11-20 15:58:58.604084] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.609 [2024-11-20 15:58:58.612058] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:00.609 [2024-11-20 15:58:58.612180] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:00.609 [2024-11-20 15:58:58.612234] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:00.609 [2024-11-20 15:58:58.612256] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:00.609 [2024-11-20 15:58:58.612271] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:00.609 [2024-11-20 15:58:58.612318] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:00.609 [2024-11-20 15:58:58.612335] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:00.609 [2024-11-20 15:58:58.612352] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.609 [2024-11-20 15:58:58.639873] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:00.609 BaseBdev1 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.609 [ 00:11:00.609 { 00:11:00.609 "name": "BaseBdev1", 00:11:00.609 "aliases": [ 00:11:00.609 "d5e55276-f78c-4895-ba4b-8b5128620ba5" 00:11:00.609 ], 00:11:00.609 "product_name": "Malloc disk", 00:11:00.609 "block_size": 512, 00:11:00.609 "num_blocks": 65536, 00:11:00.609 "uuid": "d5e55276-f78c-4895-ba4b-8b5128620ba5", 00:11:00.609 "assigned_rate_limits": { 00:11:00.609 "rw_ios_per_sec": 0, 00:11:00.609 "rw_mbytes_per_sec": 0, 00:11:00.609 "r_mbytes_per_sec": 0, 00:11:00.609 "w_mbytes_per_sec": 0 00:11:00.609 }, 00:11:00.609 "claimed": true, 00:11:00.609 "claim_type": "exclusive_write", 00:11:00.609 "zoned": false, 00:11:00.609 "supported_io_types": { 00:11:00.609 "read": true, 00:11:00.609 "write": true, 00:11:00.609 "unmap": true, 00:11:00.609 "flush": true, 00:11:00.609 "reset": true, 00:11:00.609 "nvme_admin": false, 00:11:00.609 "nvme_io": false, 00:11:00.609 "nvme_io_md": false, 00:11:00.609 "write_zeroes": true, 00:11:00.609 "zcopy": true, 00:11:00.609 "get_zone_info": false, 00:11:00.609 "zone_management": false, 00:11:00.609 "zone_append": false, 00:11:00.609 "compare": false, 00:11:00.609 "compare_and_write": false, 00:11:00.609 "abort": true, 00:11:00.609 "seek_hole": false, 00:11:00.609 "seek_data": false, 00:11:00.609 "copy": true, 00:11:00.609 "nvme_iov_md": false 00:11:00.609 }, 00:11:00.609 "memory_domains": [ 00:11:00.609 { 00:11:00.609 "dma_device_id": "system", 00:11:00.609 "dma_device_type": 1 00:11:00.609 }, 00:11:00.609 { 00:11:00.609 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:00.609 "dma_device_type": 2 00:11:00.609 } 00:11:00.609 ], 00:11:00.609 "driver_specific": {} 00:11:00.609 } 00:11:00.609 ] 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.609 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:00.609 "name": "Existed_Raid", 00:11:00.609 "uuid": "be798fe4-7a66-46fc-9ec0-90befa2b75f5", 00:11:00.609 "strip_size_kb": 64, 00:11:00.609 "state": "configuring", 00:11:00.609 "raid_level": "raid0", 00:11:00.609 "superblock": true, 00:11:00.609 "num_base_bdevs": 4, 00:11:00.609 "num_base_bdevs_discovered": 1, 00:11:00.609 "num_base_bdevs_operational": 4, 00:11:00.609 "base_bdevs_list": [ 00:11:00.609 { 00:11:00.609 "name": "BaseBdev1", 00:11:00.609 "uuid": "d5e55276-f78c-4895-ba4b-8b5128620ba5", 00:11:00.609 "is_configured": true, 00:11:00.609 "data_offset": 2048, 00:11:00.609 "data_size": 63488 00:11:00.609 }, 00:11:00.609 { 00:11:00.609 "name": "BaseBdev2", 00:11:00.609 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.609 "is_configured": false, 00:11:00.609 "data_offset": 0, 00:11:00.609 "data_size": 0 00:11:00.609 }, 00:11:00.609 { 00:11:00.609 "name": "BaseBdev3", 00:11:00.609 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.609 "is_configured": false, 00:11:00.610 "data_offset": 0, 00:11:00.610 "data_size": 0 00:11:00.610 }, 00:11:00.610 { 00:11:00.610 "name": "BaseBdev4", 00:11:00.610 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.610 "is_configured": false, 00:11:00.610 "data_offset": 0, 00:11:00.610 "data_size": 0 00:11:00.610 } 00:11:00.610 ] 00:11:00.610 }' 00:11:00.610 15:58:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:00.610 15:58:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.867 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:00.867 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.867 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.867 [2024-11-20 15:58:59.027996] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:00.867 [2024-11-20 15:58:59.028122] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:11:00.867 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.867 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:00.867 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.867 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.867 [2024-11-20 15:58:59.036043] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:00.867 [2024-11-20 15:58:59.037660] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:00.867 [2024-11-20 15:58:59.037704] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:00.868 [2024-11-20 15:58:59.037712] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:00.868 [2024-11-20 15:58:59.037722] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:00.868 [2024-11-20 15:58:59.037728] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:00.868 [2024-11-20 15:58:59.037735] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:00.868 "name": "Existed_Raid", 00:11:00.868 "uuid": "0188ce7e-8819-43b3-a3da-0c59467e3e49", 00:11:00.868 "strip_size_kb": 64, 00:11:00.868 "state": "configuring", 00:11:00.868 "raid_level": "raid0", 00:11:00.868 "superblock": true, 00:11:00.868 "num_base_bdevs": 4, 00:11:00.868 "num_base_bdevs_discovered": 1, 00:11:00.868 "num_base_bdevs_operational": 4, 00:11:00.868 "base_bdevs_list": [ 00:11:00.868 { 00:11:00.868 "name": "BaseBdev1", 00:11:00.868 "uuid": "d5e55276-f78c-4895-ba4b-8b5128620ba5", 00:11:00.868 "is_configured": true, 00:11:00.868 "data_offset": 2048, 00:11:00.868 "data_size": 63488 00:11:00.868 }, 00:11:00.868 { 00:11:00.868 "name": "BaseBdev2", 00:11:00.868 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.868 "is_configured": false, 00:11:00.868 "data_offset": 0, 00:11:00.868 "data_size": 0 00:11:00.868 }, 00:11:00.868 { 00:11:00.868 "name": "BaseBdev3", 00:11:00.868 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.868 "is_configured": false, 00:11:00.868 "data_offset": 0, 00:11:00.868 "data_size": 0 00:11:00.868 }, 00:11:00.868 { 00:11:00.868 "name": "BaseBdev4", 00:11:00.868 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.868 "is_configured": false, 00:11:00.868 "data_offset": 0, 00:11:00.868 "data_size": 0 00:11:00.868 } 00:11:00.868 ] 00:11:00.868 }' 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:00.868 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.124 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:01.124 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.124 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.379 [2024-11-20 15:58:59.386688] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:01.379 BaseBdev2 00:11:01.379 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.379 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:11:01.379 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:11:01.379 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:01.379 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:01.379 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.380 [ 00:11:01.380 { 00:11:01.380 "name": "BaseBdev2", 00:11:01.380 "aliases": [ 00:11:01.380 "12ed5260-52c4-4086-9778-d849d08ce8b5" 00:11:01.380 ], 00:11:01.380 "product_name": "Malloc disk", 00:11:01.380 "block_size": 512, 00:11:01.380 "num_blocks": 65536, 00:11:01.380 "uuid": "12ed5260-52c4-4086-9778-d849d08ce8b5", 00:11:01.380 "assigned_rate_limits": { 00:11:01.380 "rw_ios_per_sec": 0, 00:11:01.380 "rw_mbytes_per_sec": 0, 00:11:01.380 "r_mbytes_per_sec": 0, 00:11:01.380 "w_mbytes_per_sec": 0 00:11:01.380 }, 00:11:01.380 "claimed": true, 00:11:01.380 "claim_type": "exclusive_write", 00:11:01.380 "zoned": false, 00:11:01.380 "supported_io_types": { 00:11:01.380 "read": true, 00:11:01.380 "write": true, 00:11:01.380 "unmap": true, 00:11:01.380 "flush": true, 00:11:01.380 "reset": true, 00:11:01.380 "nvme_admin": false, 00:11:01.380 "nvme_io": false, 00:11:01.380 "nvme_io_md": false, 00:11:01.380 "write_zeroes": true, 00:11:01.380 "zcopy": true, 00:11:01.380 "get_zone_info": false, 00:11:01.380 "zone_management": false, 00:11:01.380 "zone_append": false, 00:11:01.380 "compare": false, 00:11:01.380 "compare_and_write": false, 00:11:01.380 "abort": true, 00:11:01.380 "seek_hole": false, 00:11:01.380 "seek_data": false, 00:11:01.380 "copy": true, 00:11:01.380 "nvme_iov_md": false 00:11:01.380 }, 00:11:01.380 "memory_domains": [ 00:11:01.380 { 00:11:01.380 "dma_device_id": "system", 00:11:01.380 "dma_device_type": 1 00:11:01.380 }, 00:11:01.380 { 00:11:01.380 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:01.380 "dma_device_type": 2 00:11:01.380 } 00:11:01.380 ], 00:11:01.380 "driver_specific": {} 00:11:01.380 } 00:11:01.380 ] 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:01.380 "name": "Existed_Raid", 00:11:01.380 "uuid": "0188ce7e-8819-43b3-a3da-0c59467e3e49", 00:11:01.380 "strip_size_kb": 64, 00:11:01.380 "state": "configuring", 00:11:01.380 "raid_level": "raid0", 00:11:01.380 "superblock": true, 00:11:01.380 "num_base_bdevs": 4, 00:11:01.380 "num_base_bdevs_discovered": 2, 00:11:01.380 "num_base_bdevs_operational": 4, 00:11:01.380 "base_bdevs_list": [ 00:11:01.380 { 00:11:01.380 "name": "BaseBdev1", 00:11:01.380 "uuid": "d5e55276-f78c-4895-ba4b-8b5128620ba5", 00:11:01.380 "is_configured": true, 00:11:01.380 "data_offset": 2048, 00:11:01.380 "data_size": 63488 00:11:01.380 }, 00:11:01.380 { 00:11:01.380 "name": "BaseBdev2", 00:11:01.380 "uuid": "12ed5260-52c4-4086-9778-d849d08ce8b5", 00:11:01.380 "is_configured": true, 00:11:01.380 "data_offset": 2048, 00:11:01.380 "data_size": 63488 00:11:01.380 }, 00:11:01.380 { 00:11:01.380 "name": "BaseBdev3", 00:11:01.380 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.380 "is_configured": false, 00:11:01.380 "data_offset": 0, 00:11:01.380 "data_size": 0 00:11:01.380 }, 00:11:01.380 { 00:11:01.380 "name": "BaseBdev4", 00:11:01.380 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.380 "is_configured": false, 00:11:01.380 "data_offset": 0, 00:11:01.380 "data_size": 0 00:11:01.380 } 00:11:01.380 ] 00:11:01.380 }' 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:01.380 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.637 [2024-11-20 15:58:59.772501] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:01.637 BaseBdev3 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.637 [ 00:11:01.637 { 00:11:01.637 "name": "BaseBdev3", 00:11:01.637 "aliases": [ 00:11:01.637 "ae84be97-d504-463e-9e9e-50a656b69b3f" 00:11:01.637 ], 00:11:01.637 "product_name": "Malloc disk", 00:11:01.637 "block_size": 512, 00:11:01.637 "num_blocks": 65536, 00:11:01.637 "uuid": "ae84be97-d504-463e-9e9e-50a656b69b3f", 00:11:01.637 "assigned_rate_limits": { 00:11:01.637 "rw_ios_per_sec": 0, 00:11:01.637 "rw_mbytes_per_sec": 0, 00:11:01.637 "r_mbytes_per_sec": 0, 00:11:01.637 "w_mbytes_per_sec": 0 00:11:01.637 }, 00:11:01.637 "claimed": true, 00:11:01.637 "claim_type": "exclusive_write", 00:11:01.637 "zoned": false, 00:11:01.637 "supported_io_types": { 00:11:01.637 "read": true, 00:11:01.637 "write": true, 00:11:01.637 "unmap": true, 00:11:01.637 "flush": true, 00:11:01.637 "reset": true, 00:11:01.637 "nvme_admin": false, 00:11:01.637 "nvme_io": false, 00:11:01.637 "nvme_io_md": false, 00:11:01.637 "write_zeroes": true, 00:11:01.637 "zcopy": true, 00:11:01.637 "get_zone_info": false, 00:11:01.637 "zone_management": false, 00:11:01.637 "zone_append": false, 00:11:01.637 "compare": false, 00:11:01.637 "compare_and_write": false, 00:11:01.637 "abort": true, 00:11:01.637 "seek_hole": false, 00:11:01.637 "seek_data": false, 00:11:01.637 "copy": true, 00:11:01.637 "nvme_iov_md": false 00:11:01.637 }, 00:11:01.637 "memory_domains": [ 00:11:01.637 { 00:11:01.637 "dma_device_id": "system", 00:11:01.637 "dma_device_type": 1 00:11:01.637 }, 00:11:01.637 { 00:11:01.637 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:01.637 "dma_device_type": 2 00:11:01.637 } 00:11:01.637 ], 00:11:01.637 "driver_specific": {} 00:11:01.637 } 00:11:01.637 ] 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.637 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:01.638 "name": "Existed_Raid", 00:11:01.638 "uuid": "0188ce7e-8819-43b3-a3da-0c59467e3e49", 00:11:01.638 "strip_size_kb": 64, 00:11:01.638 "state": "configuring", 00:11:01.638 "raid_level": "raid0", 00:11:01.638 "superblock": true, 00:11:01.638 "num_base_bdevs": 4, 00:11:01.638 "num_base_bdevs_discovered": 3, 00:11:01.638 "num_base_bdevs_operational": 4, 00:11:01.638 "base_bdevs_list": [ 00:11:01.638 { 00:11:01.638 "name": "BaseBdev1", 00:11:01.638 "uuid": "d5e55276-f78c-4895-ba4b-8b5128620ba5", 00:11:01.638 "is_configured": true, 00:11:01.638 "data_offset": 2048, 00:11:01.638 "data_size": 63488 00:11:01.638 }, 00:11:01.638 { 00:11:01.638 "name": "BaseBdev2", 00:11:01.638 "uuid": "12ed5260-52c4-4086-9778-d849d08ce8b5", 00:11:01.638 "is_configured": true, 00:11:01.638 "data_offset": 2048, 00:11:01.638 "data_size": 63488 00:11:01.638 }, 00:11:01.638 { 00:11:01.638 "name": "BaseBdev3", 00:11:01.638 "uuid": "ae84be97-d504-463e-9e9e-50a656b69b3f", 00:11:01.638 "is_configured": true, 00:11:01.638 "data_offset": 2048, 00:11:01.638 "data_size": 63488 00:11:01.638 }, 00:11:01.638 { 00:11:01.638 "name": "BaseBdev4", 00:11:01.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.638 "is_configured": false, 00:11:01.638 "data_offset": 0, 00:11:01.638 "data_size": 0 00:11:01.638 } 00:11:01.638 ] 00:11:01.638 }' 00:11:01.638 15:58:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:01.638 15:58:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.895 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:01.895 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.895 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.895 [2024-11-20 15:59:00.119200] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:01.895 [2024-11-20 15:59:00.119400] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:11:01.895 [2024-11-20 15:59:00.119412] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:01.895 [2024-11-20 15:59:00.119632] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:11:01.895 BaseBdev4 00:11:01.895 [2024-11-20 15:59:00.119781] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:11:01.895 [2024-11-20 15:59:00.119791] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:11:01.895 [2024-11-20 15:59:00.119899] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:01.895 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.895 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:11:01.895 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:11:01.895 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:01.895 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:01.895 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:01.895 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:01.895 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:01.895 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.895 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.895 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.895 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:01.895 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.895 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.895 [ 00:11:01.895 { 00:11:01.895 "name": "BaseBdev4", 00:11:01.895 "aliases": [ 00:11:01.895 "ab082487-11aa-4410-aafd-5ab83a26a732" 00:11:01.895 ], 00:11:01.895 "product_name": "Malloc disk", 00:11:01.895 "block_size": 512, 00:11:01.895 "num_blocks": 65536, 00:11:01.895 "uuid": "ab082487-11aa-4410-aafd-5ab83a26a732", 00:11:01.895 "assigned_rate_limits": { 00:11:01.895 "rw_ios_per_sec": 0, 00:11:01.895 "rw_mbytes_per_sec": 0, 00:11:01.895 "r_mbytes_per_sec": 0, 00:11:01.895 "w_mbytes_per_sec": 0 00:11:01.895 }, 00:11:01.895 "claimed": true, 00:11:01.895 "claim_type": "exclusive_write", 00:11:01.895 "zoned": false, 00:11:01.895 "supported_io_types": { 00:11:01.895 "read": true, 00:11:01.895 "write": true, 00:11:01.895 "unmap": true, 00:11:01.895 "flush": true, 00:11:01.895 "reset": true, 00:11:01.895 "nvme_admin": false, 00:11:01.895 "nvme_io": false, 00:11:01.895 "nvme_io_md": false, 00:11:01.895 "write_zeroes": true, 00:11:01.895 "zcopy": true, 00:11:01.895 "get_zone_info": false, 00:11:01.895 "zone_management": false, 00:11:01.895 "zone_append": false, 00:11:01.895 "compare": false, 00:11:01.895 "compare_and_write": false, 00:11:01.895 "abort": true, 00:11:01.895 "seek_hole": false, 00:11:01.895 "seek_data": false, 00:11:01.895 "copy": true, 00:11:01.895 "nvme_iov_md": false 00:11:01.895 }, 00:11:01.895 "memory_domains": [ 00:11:01.895 { 00:11:01.895 "dma_device_id": "system", 00:11:01.895 "dma_device_type": 1 00:11:01.895 }, 00:11:01.895 { 00:11:01.895 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:01.895 "dma_device_type": 2 00:11:01.895 } 00:11:01.895 ], 00:11:01.895 "driver_specific": {} 00:11:01.895 } 00:11:01.895 ] 00:11:02.152 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.152 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:02.152 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:02.152 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:02.153 "name": "Existed_Raid", 00:11:02.153 "uuid": "0188ce7e-8819-43b3-a3da-0c59467e3e49", 00:11:02.153 "strip_size_kb": 64, 00:11:02.153 "state": "online", 00:11:02.153 "raid_level": "raid0", 00:11:02.153 "superblock": true, 00:11:02.153 "num_base_bdevs": 4, 00:11:02.153 "num_base_bdevs_discovered": 4, 00:11:02.153 "num_base_bdevs_operational": 4, 00:11:02.153 "base_bdevs_list": [ 00:11:02.153 { 00:11:02.153 "name": "BaseBdev1", 00:11:02.153 "uuid": "d5e55276-f78c-4895-ba4b-8b5128620ba5", 00:11:02.153 "is_configured": true, 00:11:02.153 "data_offset": 2048, 00:11:02.153 "data_size": 63488 00:11:02.153 }, 00:11:02.153 { 00:11:02.153 "name": "BaseBdev2", 00:11:02.153 "uuid": "12ed5260-52c4-4086-9778-d849d08ce8b5", 00:11:02.153 "is_configured": true, 00:11:02.153 "data_offset": 2048, 00:11:02.153 "data_size": 63488 00:11:02.153 }, 00:11:02.153 { 00:11:02.153 "name": "BaseBdev3", 00:11:02.153 "uuid": "ae84be97-d504-463e-9e9e-50a656b69b3f", 00:11:02.153 "is_configured": true, 00:11:02.153 "data_offset": 2048, 00:11:02.153 "data_size": 63488 00:11:02.153 }, 00:11:02.153 { 00:11:02.153 "name": "BaseBdev4", 00:11:02.153 "uuid": "ab082487-11aa-4410-aafd-5ab83a26a732", 00:11:02.153 "is_configured": true, 00:11:02.153 "data_offset": 2048, 00:11:02.153 "data_size": 63488 00:11:02.153 } 00:11:02.153 ] 00:11:02.153 }' 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:02.153 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.410 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:11:02.410 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:02.410 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:02.410 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:02.410 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:11:02.410 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:02.410 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:02.410 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:02.410 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.410 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.410 [2024-11-20 15:59:00.459613] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:02.410 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.410 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:02.410 "name": "Existed_Raid", 00:11:02.410 "aliases": [ 00:11:02.410 "0188ce7e-8819-43b3-a3da-0c59467e3e49" 00:11:02.410 ], 00:11:02.410 "product_name": "Raid Volume", 00:11:02.410 "block_size": 512, 00:11:02.410 "num_blocks": 253952, 00:11:02.410 "uuid": "0188ce7e-8819-43b3-a3da-0c59467e3e49", 00:11:02.410 "assigned_rate_limits": { 00:11:02.410 "rw_ios_per_sec": 0, 00:11:02.410 "rw_mbytes_per_sec": 0, 00:11:02.410 "r_mbytes_per_sec": 0, 00:11:02.410 "w_mbytes_per_sec": 0 00:11:02.410 }, 00:11:02.410 "claimed": false, 00:11:02.410 "zoned": false, 00:11:02.410 "supported_io_types": { 00:11:02.410 "read": true, 00:11:02.410 "write": true, 00:11:02.410 "unmap": true, 00:11:02.410 "flush": true, 00:11:02.410 "reset": true, 00:11:02.410 "nvme_admin": false, 00:11:02.410 "nvme_io": false, 00:11:02.410 "nvme_io_md": false, 00:11:02.410 "write_zeroes": true, 00:11:02.410 "zcopy": false, 00:11:02.410 "get_zone_info": false, 00:11:02.410 "zone_management": false, 00:11:02.410 "zone_append": false, 00:11:02.410 "compare": false, 00:11:02.410 "compare_and_write": false, 00:11:02.410 "abort": false, 00:11:02.410 "seek_hole": false, 00:11:02.410 "seek_data": false, 00:11:02.410 "copy": false, 00:11:02.410 "nvme_iov_md": false 00:11:02.410 }, 00:11:02.410 "memory_domains": [ 00:11:02.410 { 00:11:02.410 "dma_device_id": "system", 00:11:02.410 "dma_device_type": 1 00:11:02.411 }, 00:11:02.411 { 00:11:02.411 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:02.411 "dma_device_type": 2 00:11:02.411 }, 00:11:02.411 { 00:11:02.411 "dma_device_id": "system", 00:11:02.411 "dma_device_type": 1 00:11:02.411 }, 00:11:02.411 { 00:11:02.411 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:02.411 "dma_device_type": 2 00:11:02.411 }, 00:11:02.411 { 00:11:02.411 "dma_device_id": "system", 00:11:02.411 "dma_device_type": 1 00:11:02.411 }, 00:11:02.411 { 00:11:02.411 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:02.411 "dma_device_type": 2 00:11:02.411 }, 00:11:02.411 { 00:11:02.411 "dma_device_id": "system", 00:11:02.411 "dma_device_type": 1 00:11:02.411 }, 00:11:02.411 { 00:11:02.411 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:02.411 "dma_device_type": 2 00:11:02.411 } 00:11:02.411 ], 00:11:02.411 "driver_specific": { 00:11:02.411 "raid": { 00:11:02.411 "uuid": "0188ce7e-8819-43b3-a3da-0c59467e3e49", 00:11:02.411 "strip_size_kb": 64, 00:11:02.411 "state": "online", 00:11:02.411 "raid_level": "raid0", 00:11:02.411 "superblock": true, 00:11:02.411 "num_base_bdevs": 4, 00:11:02.411 "num_base_bdevs_discovered": 4, 00:11:02.411 "num_base_bdevs_operational": 4, 00:11:02.411 "base_bdevs_list": [ 00:11:02.411 { 00:11:02.411 "name": "BaseBdev1", 00:11:02.411 "uuid": "d5e55276-f78c-4895-ba4b-8b5128620ba5", 00:11:02.411 "is_configured": true, 00:11:02.411 "data_offset": 2048, 00:11:02.411 "data_size": 63488 00:11:02.411 }, 00:11:02.411 { 00:11:02.411 "name": "BaseBdev2", 00:11:02.411 "uuid": "12ed5260-52c4-4086-9778-d849d08ce8b5", 00:11:02.411 "is_configured": true, 00:11:02.411 "data_offset": 2048, 00:11:02.411 "data_size": 63488 00:11:02.411 }, 00:11:02.411 { 00:11:02.411 "name": "BaseBdev3", 00:11:02.411 "uuid": "ae84be97-d504-463e-9e9e-50a656b69b3f", 00:11:02.411 "is_configured": true, 00:11:02.411 "data_offset": 2048, 00:11:02.411 "data_size": 63488 00:11:02.411 }, 00:11:02.411 { 00:11:02.411 "name": "BaseBdev4", 00:11:02.411 "uuid": "ab082487-11aa-4410-aafd-5ab83a26a732", 00:11:02.411 "is_configured": true, 00:11:02.411 "data_offset": 2048, 00:11:02.411 "data_size": 63488 00:11:02.411 } 00:11:02.411 ] 00:11:02.411 } 00:11:02.411 } 00:11:02.411 }' 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:11:02.411 BaseBdev2 00:11:02.411 BaseBdev3 00:11:02.411 BaseBdev4' 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.411 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.669 [2024-11-20 15:59:00.691396] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:02.669 [2024-11-20 15:59:00.691493] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:02.669 [2024-11-20 15:59:00.691574] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:02.669 "name": "Existed_Raid", 00:11:02.669 "uuid": "0188ce7e-8819-43b3-a3da-0c59467e3e49", 00:11:02.669 "strip_size_kb": 64, 00:11:02.669 "state": "offline", 00:11:02.669 "raid_level": "raid0", 00:11:02.669 "superblock": true, 00:11:02.669 "num_base_bdevs": 4, 00:11:02.669 "num_base_bdevs_discovered": 3, 00:11:02.669 "num_base_bdevs_operational": 3, 00:11:02.669 "base_bdevs_list": [ 00:11:02.669 { 00:11:02.669 "name": null, 00:11:02.669 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:02.669 "is_configured": false, 00:11:02.669 "data_offset": 0, 00:11:02.669 "data_size": 63488 00:11:02.669 }, 00:11:02.669 { 00:11:02.669 "name": "BaseBdev2", 00:11:02.669 "uuid": "12ed5260-52c4-4086-9778-d849d08ce8b5", 00:11:02.669 "is_configured": true, 00:11:02.669 "data_offset": 2048, 00:11:02.669 "data_size": 63488 00:11:02.669 }, 00:11:02.669 { 00:11:02.669 "name": "BaseBdev3", 00:11:02.669 "uuid": "ae84be97-d504-463e-9e9e-50a656b69b3f", 00:11:02.669 "is_configured": true, 00:11:02.669 "data_offset": 2048, 00:11:02.669 "data_size": 63488 00:11:02.669 }, 00:11:02.669 { 00:11:02.669 "name": "BaseBdev4", 00:11:02.669 "uuid": "ab082487-11aa-4410-aafd-5ab83a26a732", 00:11:02.669 "is_configured": true, 00:11:02.669 "data_offset": 2048, 00:11:02.669 "data_size": 63488 00:11:02.669 } 00:11:02.669 ] 00:11:02.669 }' 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:02.669 15:59:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.926 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.927 [2024-11-20 15:59:01.082195] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.927 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.927 [2024-11-20 15:59:01.164351] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.184 [2024-11-20 15:59:01.247748] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:11:03.184 [2024-11-20 15:59:01.247866] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.184 BaseBdev2 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.184 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.184 [ 00:11:03.184 { 00:11:03.184 "name": "BaseBdev2", 00:11:03.184 "aliases": [ 00:11:03.184 "138edb0f-09f9-4904-a9d1-f7bf441be23c" 00:11:03.184 ], 00:11:03.184 "product_name": "Malloc disk", 00:11:03.184 "block_size": 512, 00:11:03.184 "num_blocks": 65536, 00:11:03.184 "uuid": "138edb0f-09f9-4904-a9d1-f7bf441be23c", 00:11:03.184 "assigned_rate_limits": { 00:11:03.184 "rw_ios_per_sec": 0, 00:11:03.184 "rw_mbytes_per_sec": 0, 00:11:03.184 "r_mbytes_per_sec": 0, 00:11:03.184 "w_mbytes_per_sec": 0 00:11:03.184 }, 00:11:03.184 "claimed": false, 00:11:03.184 "zoned": false, 00:11:03.184 "supported_io_types": { 00:11:03.184 "read": true, 00:11:03.184 "write": true, 00:11:03.184 "unmap": true, 00:11:03.185 "flush": true, 00:11:03.185 "reset": true, 00:11:03.185 "nvme_admin": false, 00:11:03.185 "nvme_io": false, 00:11:03.185 "nvme_io_md": false, 00:11:03.185 "write_zeroes": true, 00:11:03.185 "zcopy": true, 00:11:03.185 "get_zone_info": false, 00:11:03.185 "zone_management": false, 00:11:03.185 "zone_append": false, 00:11:03.185 "compare": false, 00:11:03.185 "compare_and_write": false, 00:11:03.185 "abort": true, 00:11:03.185 "seek_hole": false, 00:11:03.185 "seek_data": false, 00:11:03.185 "copy": true, 00:11:03.185 "nvme_iov_md": false 00:11:03.185 }, 00:11:03.185 "memory_domains": [ 00:11:03.185 { 00:11:03.185 "dma_device_id": "system", 00:11:03.185 "dma_device_type": 1 00:11:03.185 }, 00:11:03.185 { 00:11:03.185 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:03.185 "dma_device_type": 2 00:11:03.185 } 00:11:03.185 ], 00:11:03.185 "driver_specific": {} 00:11:03.185 } 00:11:03.185 ] 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.185 BaseBdev3 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.185 [ 00:11:03.185 { 00:11:03.185 "name": "BaseBdev3", 00:11:03.185 "aliases": [ 00:11:03.185 "21ce5b44-9ce4-49f4-b3e3-9938b98945cd" 00:11:03.185 ], 00:11:03.185 "product_name": "Malloc disk", 00:11:03.185 "block_size": 512, 00:11:03.185 "num_blocks": 65536, 00:11:03.185 "uuid": "21ce5b44-9ce4-49f4-b3e3-9938b98945cd", 00:11:03.185 "assigned_rate_limits": { 00:11:03.185 "rw_ios_per_sec": 0, 00:11:03.185 "rw_mbytes_per_sec": 0, 00:11:03.185 "r_mbytes_per_sec": 0, 00:11:03.185 "w_mbytes_per_sec": 0 00:11:03.185 }, 00:11:03.185 "claimed": false, 00:11:03.185 "zoned": false, 00:11:03.185 "supported_io_types": { 00:11:03.185 "read": true, 00:11:03.185 "write": true, 00:11:03.185 "unmap": true, 00:11:03.185 "flush": true, 00:11:03.185 "reset": true, 00:11:03.185 "nvme_admin": false, 00:11:03.185 "nvme_io": false, 00:11:03.185 "nvme_io_md": false, 00:11:03.185 "write_zeroes": true, 00:11:03.185 "zcopy": true, 00:11:03.185 "get_zone_info": false, 00:11:03.185 "zone_management": false, 00:11:03.185 "zone_append": false, 00:11:03.185 "compare": false, 00:11:03.185 "compare_and_write": false, 00:11:03.185 "abort": true, 00:11:03.185 "seek_hole": false, 00:11:03.185 "seek_data": false, 00:11:03.185 "copy": true, 00:11:03.185 "nvme_iov_md": false 00:11:03.185 }, 00:11:03.185 "memory_domains": [ 00:11:03.185 { 00:11:03.185 "dma_device_id": "system", 00:11:03.185 "dma_device_type": 1 00:11:03.185 }, 00:11:03.185 { 00:11:03.185 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:03.185 "dma_device_type": 2 00:11:03.185 } 00:11:03.185 ], 00:11:03.185 "driver_specific": {} 00:11:03.185 } 00:11:03.185 ] 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.185 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.442 BaseBdev4 00:11:03.442 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.442 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.443 [ 00:11:03.443 { 00:11:03.443 "name": "BaseBdev4", 00:11:03.443 "aliases": [ 00:11:03.443 "b0467226-c184-47ba-86b6-a04292fe0837" 00:11:03.443 ], 00:11:03.443 "product_name": "Malloc disk", 00:11:03.443 "block_size": 512, 00:11:03.443 "num_blocks": 65536, 00:11:03.443 "uuid": "b0467226-c184-47ba-86b6-a04292fe0837", 00:11:03.443 "assigned_rate_limits": { 00:11:03.443 "rw_ios_per_sec": 0, 00:11:03.443 "rw_mbytes_per_sec": 0, 00:11:03.443 "r_mbytes_per_sec": 0, 00:11:03.443 "w_mbytes_per_sec": 0 00:11:03.443 }, 00:11:03.443 "claimed": false, 00:11:03.443 "zoned": false, 00:11:03.443 "supported_io_types": { 00:11:03.443 "read": true, 00:11:03.443 "write": true, 00:11:03.443 "unmap": true, 00:11:03.443 "flush": true, 00:11:03.443 "reset": true, 00:11:03.443 "nvme_admin": false, 00:11:03.443 "nvme_io": false, 00:11:03.443 "nvme_io_md": false, 00:11:03.443 "write_zeroes": true, 00:11:03.443 "zcopy": true, 00:11:03.443 "get_zone_info": false, 00:11:03.443 "zone_management": false, 00:11:03.443 "zone_append": false, 00:11:03.443 "compare": false, 00:11:03.443 "compare_and_write": false, 00:11:03.443 "abort": true, 00:11:03.443 "seek_hole": false, 00:11:03.443 "seek_data": false, 00:11:03.443 "copy": true, 00:11:03.443 "nvme_iov_md": false 00:11:03.443 }, 00:11:03.443 "memory_domains": [ 00:11:03.443 { 00:11:03.443 "dma_device_id": "system", 00:11:03.443 "dma_device_type": 1 00:11:03.443 }, 00:11:03.443 { 00:11:03.443 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:03.443 "dma_device_type": 2 00:11:03.443 } 00:11:03.443 ], 00:11:03.443 "driver_specific": {} 00:11:03.443 } 00:11:03.443 ] 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.443 [2024-11-20 15:59:01.483392] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:03.443 [2024-11-20 15:59:01.483510] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:03.443 [2024-11-20 15:59:01.483574] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:03.443 [2024-11-20 15:59:01.485139] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:03.443 [2024-11-20 15:59:01.485270] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:03.443 "name": "Existed_Raid", 00:11:03.443 "uuid": "720cd1e7-0f40-4681-9356-1ba3d9372c53", 00:11:03.443 "strip_size_kb": 64, 00:11:03.443 "state": "configuring", 00:11:03.443 "raid_level": "raid0", 00:11:03.443 "superblock": true, 00:11:03.443 "num_base_bdevs": 4, 00:11:03.443 "num_base_bdevs_discovered": 3, 00:11:03.443 "num_base_bdevs_operational": 4, 00:11:03.443 "base_bdevs_list": [ 00:11:03.443 { 00:11:03.443 "name": "BaseBdev1", 00:11:03.443 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:03.443 "is_configured": false, 00:11:03.443 "data_offset": 0, 00:11:03.443 "data_size": 0 00:11:03.443 }, 00:11:03.443 { 00:11:03.443 "name": "BaseBdev2", 00:11:03.443 "uuid": "138edb0f-09f9-4904-a9d1-f7bf441be23c", 00:11:03.443 "is_configured": true, 00:11:03.443 "data_offset": 2048, 00:11:03.443 "data_size": 63488 00:11:03.443 }, 00:11:03.443 { 00:11:03.443 "name": "BaseBdev3", 00:11:03.443 "uuid": "21ce5b44-9ce4-49f4-b3e3-9938b98945cd", 00:11:03.443 "is_configured": true, 00:11:03.443 "data_offset": 2048, 00:11:03.443 "data_size": 63488 00:11:03.443 }, 00:11:03.443 { 00:11:03.443 "name": "BaseBdev4", 00:11:03.443 "uuid": "b0467226-c184-47ba-86b6-a04292fe0837", 00:11:03.443 "is_configured": true, 00:11:03.443 "data_offset": 2048, 00:11:03.443 "data_size": 63488 00:11:03.443 } 00:11:03.443 ] 00:11:03.443 }' 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:03.443 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.700 [2024-11-20 15:59:01.823455] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:03.700 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.701 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:03.701 "name": "Existed_Raid", 00:11:03.701 "uuid": "720cd1e7-0f40-4681-9356-1ba3d9372c53", 00:11:03.701 "strip_size_kb": 64, 00:11:03.701 "state": "configuring", 00:11:03.701 "raid_level": "raid0", 00:11:03.701 "superblock": true, 00:11:03.701 "num_base_bdevs": 4, 00:11:03.701 "num_base_bdevs_discovered": 2, 00:11:03.701 "num_base_bdevs_operational": 4, 00:11:03.701 "base_bdevs_list": [ 00:11:03.701 { 00:11:03.701 "name": "BaseBdev1", 00:11:03.701 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:03.701 "is_configured": false, 00:11:03.701 "data_offset": 0, 00:11:03.701 "data_size": 0 00:11:03.701 }, 00:11:03.701 { 00:11:03.701 "name": null, 00:11:03.701 "uuid": "138edb0f-09f9-4904-a9d1-f7bf441be23c", 00:11:03.701 "is_configured": false, 00:11:03.701 "data_offset": 0, 00:11:03.701 "data_size": 63488 00:11:03.701 }, 00:11:03.701 { 00:11:03.701 "name": "BaseBdev3", 00:11:03.701 "uuid": "21ce5b44-9ce4-49f4-b3e3-9938b98945cd", 00:11:03.701 "is_configured": true, 00:11:03.701 "data_offset": 2048, 00:11:03.701 "data_size": 63488 00:11:03.701 }, 00:11:03.701 { 00:11:03.701 "name": "BaseBdev4", 00:11:03.701 "uuid": "b0467226-c184-47ba-86b6-a04292fe0837", 00:11:03.701 "is_configured": true, 00:11:03.701 "data_offset": 2048, 00:11:03.701 "data_size": 63488 00:11:03.701 } 00:11:03.701 ] 00:11:03.701 }' 00:11:03.701 15:59:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:03.701 15:59:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.958 [2024-11-20 15:59:02.201645] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:03.958 BaseBdev1 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.958 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.215 [ 00:11:04.215 { 00:11:04.215 "name": "BaseBdev1", 00:11:04.215 "aliases": [ 00:11:04.215 "7100ef5e-50fe-4c43-9621-173e37244318" 00:11:04.215 ], 00:11:04.215 "product_name": "Malloc disk", 00:11:04.215 "block_size": 512, 00:11:04.215 "num_blocks": 65536, 00:11:04.215 "uuid": "7100ef5e-50fe-4c43-9621-173e37244318", 00:11:04.215 "assigned_rate_limits": { 00:11:04.215 "rw_ios_per_sec": 0, 00:11:04.215 "rw_mbytes_per_sec": 0, 00:11:04.215 "r_mbytes_per_sec": 0, 00:11:04.215 "w_mbytes_per_sec": 0 00:11:04.215 }, 00:11:04.215 "claimed": true, 00:11:04.215 "claim_type": "exclusive_write", 00:11:04.215 "zoned": false, 00:11:04.215 "supported_io_types": { 00:11:04.215 "read": true, 00:11:04.215 "write": true, 00:11:04.215 "unmap": true, 00:11:04.215 "flush": true, 00:11:04.215 "reset": true, 00:11:04.215 "nvme_admin": false, 00:11:04.215 "nvme_io": false, 00:11:04.215 "nvme_io_md": false, 00:11:04.215 "write_zeroes": true, 00:11:04.215 "zcopy": true, 00:11:04.215 "get_zone_info": false, 00:11:04.215 "zone_management": false, 00:11:04.215 "zone_append": false, 00:11:04.215 "compare": false, 00:11:04.215 "compare_and_write": false, 00:11:04.215 "abort": true, 00:11:04.215 "seek_hole": false, 00:11:04.215 "seek_data": false, 00:11:04.215 "copy": true, 00:11:04.215 "nvme_iov_md": false 00:11:04.215 }, 00:11:04.215 "memory_domains": [ 00:11:04.215 { 00:11:04.215 "dma_device_id": "system", 00:11:04.215 "dma_device_type": 1 00:11:04.215 }, 00:11:04.215 { 00:11:04.215 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:04.215 "dma_device_type": 2 00:11:04.215 } 00:11:04.215 ], 00:11:04.215 "driver_specific": {} 00:11:04.215 } 00:11:04.215 ] 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.215 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:04.215 "name": "Existed_Raid", 00:11:04.215 "uuid": "720cd1e7-0f40-4681-9356-1ba3d9372c53", 00:11:04.215 "strip_size_kb": 64, 00:11:04.215 "state": "configuring", 00:11:04.215 "raid_level": "raid0", 00:11:04.215 "superblock": true, 00:11:04.215 "num_base_bdevs": 4, 00:11:04.215 "num_base_bdevs_discovered": 3, 00:11:04.215 "num_base_bdevs_operational": 4, 00:11:04.215 "base_bdevs_list": [ 00:11:04.215 { 00:11:04.215 "name": "BaseBdev1", 00:11:04.215 "uuid": "7100ef5e-50fe-4c43-9621-173e37244318", 00:11:04.215 "is_configured": true, 00:11:04.215 "data_offset": 2048, 00:11:04.215 "data_size": 63488 00:11:04.215 }, 00:11:04.215 { 00:11:04.215 "name": null, 00:11:04.215 "uuid": "138edb0f-09f9-4904-a9d1-f7bf441be23c", 00:11:04.215 "is_configured": false, 00:11:04.215 "data_offset": 0, 00:11:04.215 "data_size": 63488 00:11:04.215 }, 00:11:04.215 { 00:11:04.215 "name": "BaseBdev3", 00:11:04.215 "uuid": "21ce5b44-9ce4-49f4-b3e3-9938b98945cd", 00:11:04.215 "is_configured": true, 00:11:04.215 "data_offset": 2048, 00:11:04.215 "data_size": 63488 00:11:04.215 }, 00:11:04.215 { 00:11:04.215 "name": "BaseBdev4", 00:11:04.215 "uuid": "b0467226-c184-47ba-86b6-a04292fe0837", 00:11:04.215 "is_configured": true, 00:11:04.215 "data_offset": 2048, 00:11:04.215 "data_size": 63488 00:11:04.216 } 00:11:04.216 ] 00:11:04.216 }' 00:11:04.216 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:04.216 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.475 [2024-11-20 15:59:02.589814] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:04.475 "name": "Existed_Raid", 00:11:04.475 "uuid": "720cd1e7-0f40-4681-9356-1ba3d9372c53", 00:11:04.475 "strip_size_kb": 64, 00:11:04.475 "state": "configuring", 00:11:04.475 "raid_level": "raid0", 00:11:04.475 "superblock": true, 00:11:04.475 "num_base_bdevs": 4, 00:11:04.475 "num_base_bdevs_discovered": 2, 00:11:04.475 "num_base_bdevs_operational": 4, 00:11:04.475 "base_bdevs_list": [ 00:11:04.475 { 00:11:04.475 "name": "BaseBdev1", 00:11:04.475 "uuid": "7100ef5e-50fe-4c43-9621-173e37244318", 00:11:04.475 "is_configured": true, 00:11:04.475 "data_offset": 2048, 00:11:04.475 "data_size": 63488 00:11:04.475 }, 00:11:04.475 { 00:11:04.475 "name": null, 00:11:04.475 "uuid": "138edb0f-09f9-4904-a9d1-f7bf441be23c", 00:11:04.475 "is_configured": false, 00:11:04.475 "data_offset": 0, 00:11:04.475 "data_size": 63488 00:11:04.475 }, 00:11:04.475 { 00:11:04.475 "name": null, 00:11:04.475 "uuid": "21ce5b44-9ce4-49f4-b3e3-9938b98945cd", 00:11:04.475 "is_configured": false, 00:11:04.475 "data_offset": 0, 00:11:04.475 "data_size": 63488 00:11:04.475 }, 00:11:04.475 { 00:11:04.475 "name": "BaseBdev4", 00:11:04.475 "uuid": "b0467226-c184-47ba-86b6-a04292fe0837", 00:11:04.475 "is_configured": true, 00:11:04.475 "data_offset": 2048, 00:11:04.475 "data_size": 63488 00:11:04.475 } 00:11:04.475 ] 00:11:04.475 }' 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:04.475 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.732 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:04.732 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:04.732 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.732 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.732 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.732 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:11:04.732 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:11:04.732 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.732 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.732 [2024-11-20 15:59:02.941855] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:04.732 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.732 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:04.732 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:04.733 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:04.733 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:04.733 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:04.733 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:04.733 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:04.733 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:04.733 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:04.733 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:04.733 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:04.733 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:04.733 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.733 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.733 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.733 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:04.733 "name": "Existed_Raid", 00:11:04.733 "uuid": "720cd1e7-0f40-4681-9356-1ba3d9372c53", 00:11:04.733 "strip_size_kb": 64, 00:11:04.733 "state": "configuring", 00:11:04.733 "raid_level": "raid0", 00:11:04.733 "superblock": true, 00:11:04.733 "num_base_bdevs": 4, 00:11:04.733 "num_base_bdevs_discovered": 3, 00:11:04.733 "num_base_bdevs_operational": 4, 00:11:04.733 "base_bdevs_list": [ 00:11:04.733 { 00:11:04.733 "name": "BaseBdev1", 00:11:04.733 "uuid": "7100ef5e-50fe-4c43-9621-173e37244318", 00:11:04.733 "is_configured": true, 00:11:04.733 "data_offset": 2048, 00:11:04.733 "data_size": 63488 00:11:04.733 }, 00:11:04.733 { 00:11:04.733 "name": null, 00:11:04.733 "uuid": "138edb0f-09f9-4904-a9d1-f7bf441be23c", 00:11:04.733 "is_configured": false, 00:11:04.733 "data_offset": 0, 00:11:04.733 "data_size": 63488 00:11:04.733 }, 00:11:04.733 { 00:11:04.733 "name": "BaseBdev3", 00:11:04.733 "uuid": "21ce5b44-9ce4-49f4-b3e3-9938b98945cd", 00:11:04.733 "is_configured": true, 00:11:04.733 "data_offset": 2048, 00:11:04.733 "data_size": 63488 00:11:04.733 }, 00:11:04.733 { 00:11:04.733 "name": "BaseBdev4", 00:11:04.733 "uuid": "b0467226-c184-47ba-86b6-a04292fe0837", 00:11:04.733 "is_configured": true, 00:11:04.733 "data_offset": 2048, 00:11:04.733 "data_size": 63488 00:11:04.733 } 00:11:04.733 ] 00:11:04.733 }' 00:11:04.733 15:59:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:04.733 15:59:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.990 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:04.990 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:04.990 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.990 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.990 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.246 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:11:05.246 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:05.246 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.246 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:05.246 [2024-11-20 15:59:03.257944] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:05.246 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.246 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:05.246 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:05.246 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:05.246 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:05.247 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:05.247 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:05.247 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:05.247 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:05.247 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:05.247 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:05.247 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:05.247 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:05.247 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.247 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:05.247 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.247 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:05.247 "name": "Existed_Raid", 00:11:05.247 "uuid": "720cd1e7-0f40-4681-9356-1ba3d9372c53", 00:11:05.247 "strip_size_kb": 64, 00:11:05.247 "state": "configuring", 00:11:05.247 "raid_level": "raid0", 00:11:05.247 "superblock": true, 00:11:05.247 "num_base_bdevs": 4, 00:11:05.247 "num_base_bdevs_discovered": 2, 00:11:05.247 "num_base_bdevs_operational": 4, 00:11:05.247 "base_bdevs_list": [ 00:11:05.247 { 00:11:05.247 "name": null, 00:11:05.247 "uuid": "7100ef5e-50fe-4c43-9621-173e37244318", 00:11:05.247 "is_configured": false, 00:11:05.247 "data_offset": 0, 00:11:05.247 "data_size": 63488 00:11:05.247 }, 00:11:05.247 { 00:11:05.247 "name": null, 00:11:05.247 "uuid": "138edb0f-09f9-4904-a9d1-f7bf441be23c", 00:11:05.247 "is_configured": false, 00:11:05.247 "data_offset": 0, 00:11:05.247 "data_size": 63488 00:11:05.247 }, 00:11:05.247 { 00:11:05.247 "name": "BaseBdev3", 00:11:05.247 "uuid": "21ce5b44-9ce4-49f4-b3e3-9938b98945cd", 00:11:05.247 "is_configured": true, 00:11:05.247 "data_offset": 2048, 00:11:05.247 "data_size": 63488 00:11:05.247 }, 00:11:05.247 { 00:11:05.247 "name": "BaseBdev4", 00:11:05.247 "uuid": "b0467226-c184-47ba-86b6-a04292fe0837", 00:11:05.247 "is_configured": true, 00:11:05.247 "data_offset": 2048, 00:11:05.247 "data_size": 63488 00:11:05.247 } 00:11:05.247 ] 00:11:05.247 }' 00:11:05.247 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:05.247 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:05.503 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:05.503 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:05.503 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.503 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:05.503 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:05.504 [2024-11-20 15:59:03.672569] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:05.504 "name": "Existed_Raid", 00:11:05.504 "uuid": "720cd1e7-0f40-4681-9356-1ba3d9372c53", 00:11:05.504 "strip_size_kb": 64, 00:11:05.504 "state": "configuring", 00:11:05.504 "raid_level": "raid0", 00:11:05.504 "superblock": true, 00:11:05.504 "num_base_bdevs": 4, 00:11:05.504 "num_base_bdevs_discovered": 3, 00:11:05.504 "num_base_bdevs_operational": 4, 00:11:05.504 "base_bdevs_list": [ 00:11:05.504 { 00:11:05.504 "name": null, 00:11:05.504 "uuid": "7100ef5e-50fe-4c43-9621-173e37244318", 00:11:05.504 "is_configured": false, 00:11:05.504 "data_offset": 0, 00:11:05.504 "data_size": 63488 00:11:05.504 }, 00:11:05.504 { 00:11:05.504 "name": "BaseBdev2", 00:11:05.504 "uuid": "138edb0f-09f9-4904-a9d1-f7bf441be23c", 00:11:05.504 "is_configured": true, 00:11:05.504 "data_offset": 2048, 00:11:05.504 "data_size": 63488 00:11:05.504 }, 00:11:05.504 { 00:11:05.504 "name": "BaseBdev3", 00:11:05.504 "uuid": "21ce5b44-9ce4-49f4-b3e3-9938b98945cd", 00:11:05.504 "is_configured": true, 00:11:05.504 "data_offset": 2048, 00:11:05.504 "data_size": 63488 00:11:05.504 }, 00:11:05.504 { 00:11:05.504 "name": "BaseBdev4", 00:11:05.504 "uuid": "b0467226-c184-47ba-86b6-a04292fe0837", 00:11:05.504 "is_configured": true, 00:11:05.504 "data_offset": 2048, 00:11:05.504 "data_size": 63488 00:11:05.504 } 00:11:05.504 ] 00:11:05.504 }' 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:05.504 15:59:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:05.761 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:05.761 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:05.761 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.761 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 7100ef5e-50fe-4c43-9621-173e37244318 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.018 NewBaseBdev 00:11:06.018 [2024-11-20 15:59:04.082774] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:11:06.018 [2024-11-20 15:59:04.082947] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:11:06.018 [2024-11-20 15:59:04.082957] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:06.018 [2024-11-20 15:59:04.083161] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:11:06.018 [2024-11-20 15:59:04.083259] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:11:06.018 [2024-11-20 15:59:04.083268] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:11:06.018 [2024-11-20 15:59:04.083357] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.018 [ 00:11:06.018 { 00:11:06.018 "name": "NewBaseBdev", 00:11:06.018 "aliases": [ 00:11:06.018 "7100ef5e-50fe-4c43-9621-173e37244318" 00:11:06.018 ], 00:11:06.018 "product_name": "Malloc disk", 00:11:06.018 "block_size": 512, 00:11:06.018 "num_blocks": 65536, 00:11:06.018 "uuid": "7100ef5e-50fe-4c43-9621-173e37244318", 00:11:06.018 "assigned_rate_limits": { 00:11:06.018 "rw_ios_per_sec": 0, 00:11:06.018 "rw_mbytes_per_sec": 0, 00:11:06.018 "r_mbytes_per_sec": 0, 00:11:06.018 "w_mbytes_per_sec": 0 00:11:06.018 }, 00:11:06.018 "claimed": true, 00:11:06.018 "claim_type": "exclusive_write", 00:11:06.018 "zoned": false, 00:11:06.018 "supported_io_types": { 00:11:06.018 "read": true, 00:11:06.018 "write": true, 00:11:06.018 "unmap": true, 00:11:06.018 "flush": true, 00:11:06.018 "reset": true, 00:11:06.018 "nvme_admin": false, 00:11:06.018 "nvme_io": false, 00:11:06.018 "nvme_io_md": false, 00:11:06.018 "write_zeroes": true, 00:11:06.018 "zcopy": true, 00:11:06.018 "get_zone_info": false, 00:11:06.018 "zone_management": false, 00:11:06.018 "zone_append": false, 00:11:06.018 "compare": false, 00:11:06.018 "compare_and_write": false, 00:11:06.018 "abort": true, 00:11:06.018 "seek_hole": false, 00:11:06.018 "seek_data": false, 00:11:06.018 "copy": true, 00:11:06.018 "nvme_iov_md": false 00:11:06.018 }, 00:11:06.018 "memory_domains": [ 00:11:06.018 { 00:11:06.018 "dma_device_id": "system", 00:11:06.018 "dma_device_type": 1 00:11:06.018 }, 00:11:06.018 { 00:11:06.018 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:06.018 "dma_device_type": 2 00:11:06.018 } 00:11:06.018 ], 00:11:06.018 "driver_specific": {} 00:11:06.018 } 00:11:06.018 ] 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:06.018 "name": "Existed_Raid", 00:11:06.018 "uuid": "720cd1e7-0f40-4681-9356-1ba3d9372c53", 00:11:06.018 "strip_size_kb": 64, 00:11:06.018 "state": "online", 00:11:06.018 "raid_level": "raid0", 00:11:06.018 "superblock": true, 00:11:06.018 "num_base_bdevs": 4, 00:11:06.018 "num_base_bdevs_discovered": 4, 00:11:06.018 "num_base_bdevs_operational": 4, 00:11:06.018 "base_bdevs_list": [ 00:11:06.018 { 00:11:06.018 "name": "NewBaseBdev", 00:11:06.018 "uuid": "7100ef5e-50fe-4c43-9621-173e37244318", 00:11:06.018 "is_configured": true, 00:11:06.018 "data_offset": 2048, 00:11:06.018 "data_size": 63488 00:11:06.018 }, 00:11:06.018 { 00:11:06.018 "name": "BaseBdev2", 00:11:06.018 "uuid": "138edb0f-09f9-4904-a9d1-f7bf441be23c", 00:11:06.018 "is_configured": true, 00:11:06.018 "data_offset": 2048, 00:11:06.018 "data_size": 63488 00:11:06.018 }, 00:11:06.018 { 00:11:06.018 "name": "BaseBdev3", 00:11:06.018 "uuid": "21ce5b44-9ce4-49f4-b3e3-9938b98945cd", 00:11:06.018 "is_configured": true, 00:11:06.018 "data_offset": 2048, 00:11:06.018 "data_size": 63488 00:11:06.018 }, 00:11:06.018 { 00:11:06.018 "name": "BaseBdev4", 00:11:06.018 "uuid": "b0467226-c184-47ba-86b6-a04292fe0837", 00:11:06.018 "is_configured": true, 00:11:06.018 "data_offset": 2048, 00:11:06.018 "data_size": 63488 00:11:06.018 } 00:11:06.018 ] 00:11:06.018 }' 00:11:06.018 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:06.019 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.276 [2024-11-20 15:59:04.427190] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:06.276 "name": "Existed_Raid", 00:11:06.276 "aliases": [ 00:11:06.276 "720cd1e7-0f40-4681-9356-1ba3d9372c53" 00:11:06.276 ], 00:11:06.276 "product_name": "Raid Volume", 00:11:06.276 "block_size": 512, 00:11:06.276 "num_blocks": 253952, 00:11:06.276 "uuid": "720cd1e7-0f40-4681-9356-1ba3d9372c53", 00:11:06.276 "assigned_rate_limits": { 00:11:06.276 "rw_ios_per_sec": 0, 00:11:06.276 "rw_mbytes_per_sec": 0, 00:11:06.276 "r_mbytes_per_sec": 0, 00:11:06.276 "w_mbytes_per_sec": 0 00:11:06.276 }, 00:11:06.276 "claimed": false, 00:11:06.276 "zoned": false, 00:11:06.276 "supported_io_types": { 00:11:06.276 "read": true, 00:11:06.276 "write": true, 00:11:06.276 "unmap": true, 00:11:06.276 "flush": true, 00:11:06.276 "reset": true, 00:11:06.276 "nvme_admin": false, 00:11:06.276 "nvme_io": false, 00:11:06.276 "nvme_io_md": false, 00:11:06.276 "write_zeroes": true, 00:11:06.276 "zcopy": false, 00:11:06.276 "get_zone_info": false, 00:11:06.276 "zone_management": false, 00:11:06.276 "zone_append": false, 00:11:06.276 "compare": false, 00:11:06.276 "compare_and_write": false, 00:11:06.276 "abort": false, 00:11:06.276 "seek_hole": false, 00:11:06.276 "seek_data": false, 00:11:06.276 "copy": false, 00:11:06.276 "nvme_iov_md": false 00:11:06.276 }, 00:11:06.276 "memory_domains": [ 00:11:06.276 { 00:11:06.276 "dma_device_id": "system", 00:11:06.276 "dma_device_type": 1 00:11:06.276 }, 00:11:06.276 { 00:11:06.276 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:06.276 "dma_device_type": 2 00:11:06.276 }, 00:11:06.276 { 00:11:06.276 "dma_device_id": "system", 00:11:06.276 "dma_device_type": 1 00:11:06.276 }, 00:11:06.276 { 00:11:06.276 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:06.276 "dma_device_type": 2 00:11:06.276 }, 00:11:06.276 { 00:11:06.276 "dma_device_id": "system", 00:11:06.276 "dma_device_type": 1 00:11:06.276 }, 00:11:06.276 { 00:11:06.276 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:06.276 "dma_device_type": 2 00:11:06.276 }, 00:11:06.276 { 00:11:06.276 "dma_device_id": "system", 00:11:06.276 "dma_device_type": 1 00:11:06.276 }, 00:11:06.276 { 00:11:06.276 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:06.276 "dma_device_type": 2 00:11:06.276 } 00:11:06.276 ], 00:11:06.276 "driver_specific": { 00:11:06.276 "raid": { 00:11:06.276 "uuid": "720cd1e7-0f40-4681-9356-1ba3d9372c53", 00:11:06.276 "strip_size_kb": 64, 00:11:06.276 "state": "online", 00:11:06.276 "raid_level": "raid0", 00:11:06.276 "superblock": true, 00:11:06.276 "num_base_bdevs": 4, 00:11:06.276 "num_base_bdevs_discovered": 4, 00:11:06.276 "num_base_bdevs_operational": 4, 00:11:06.276 "base_bdevs_list": [ 00:11:06.276 { 00:11:06.276 "name": "NewBaseBdev", 00:11:06.276 "uuid": "7100ef5e-50fe-4c43-9621-173e37244318", 00:11:06.276 "is_configured": true, 00:11:06.276 "data_offset": 2048, 00:11:06.276 "data_size": 63488 00:11:06.276 }, 00:11:06.276 { 00:11:06.276 "name": "BaseBdev2", 00:11:06.276 "uuid": "138edb0f-09f9-4904-a9d1-f7bf441be23c", 00:11:06.276 "is_configured": true, 00:11:06.276 "data_offset": 2048, 00:11:06.276 "data_size": 63488 00:11:06.276 }, 00:11:06.276 { 00:11:06.276 "name": "BaseBdev3", 00:11:06.276 "uuid": "21ce5b44-9ce4-49f4-b3e3-9938b98945cd", 00:11:06.276 "is_configured": true, 00:11:06.276 "data_offset": 2048, 00:11:06.276 "data_size": 63488 00:11:06.276 }, 00:11:06.276 { 00:11:06.276 "name": "BaseBdev4", 00:11:06.276 "uuid": "b0467226-c184-47ba-86b6-a04292fe0837", 00:11:06.276 "is_configured": true, 00:11:06.276 "data_offset": 2048, 00:11:06.276 "data_size": 63488 00:11:06.276 } 00:11:06.276 ] 00:11:06.276 } 00:11:06.276 } 00:11:06.276 }' 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:11:06.276 BaseBdev2 00:11:06.276 BaseBdev3 00:11:06.276 BaseBdev4' 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:06.276 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.549 [2024-11-20 15:59:04.630924] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:06.549 [2024-11-20 15:59:04.630950] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:06.549 [2024-11-20 15:59:04.631007] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:06.549 [2024-11-20 15:59:04.631062] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:06.549 [2024-11-20 15:59:04.631070] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 68201 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 68201 ']' 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 68201 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 68201 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:06.549 killing process with pid 68201 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 68201' 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 68201 00:11:06.549 [2024-11-20 15:59:04.662870] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:06.549 15:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 68201 00:11:06.810 [2024-11-20 15:59:04.867768] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:07.418 15:59:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:11:07.418 00:11:07.418 real 0m8.244s 00:11:07.418 user 0m13.194s 00:11:07.418 sys 0m1.350s 00:11:07.418 ************************************ 00:11:07.418 END TEST raid_state_function_test_sb 00:11:07.418 ************************************ 00:11:07.418 15:59:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:07.418 15:59:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:07.686 15:59:05 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 4 00:11:07.686 15:59:05 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:11:07.686 15:59:05 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:07.686 15:59:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:07.686 ************************************ 00:11:07.686 START TEST raid_superblock_test 00:11:07.686 ************************************ 00:11:07.686 15:59:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 4 00:11:07.686 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:11:07.686 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=68838 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 68838 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 68838 ']' 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:07.687 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:07.687 15:59:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.687 [2024-11-20 15:59:05.730447] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:11:07.687 [2024-11-20 15:59:05.730565] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid68838 ] 00:11:07.687 [2024-11-20 15:59:05.888623] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:07.945 [2024-11-20 15:59:05.995005] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:07.945 [2024-11-20 15:59:06.137729] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:07.945 [2024-11-20 15:59:06.137787] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.513 malloc1 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.513 [2024-11-20 15:59:06.609608] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:11:08.513 [2024-11-20 15:59:06.609663] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:08.513 [2024-11-20 15:59:06.609696] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:08.513 [2024-11-20 15:59:06.609706] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:08.513 [2024-11-20 15:59:06.611859] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:08.513 [2024-11-20 15:59:06.611888] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:11:08.513 pt1 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.513 malloc2 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.513 [2024-11-20 15:59:06.646581] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:08.513 [2024-11-20 15:59:06.646624] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:08.513 [2024-11-20 15:59:06.646646] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:08.513 [2024-11-20 15:59:06.646654] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:08.513 [2024-11-20 15:59:06.648796] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:08.513 [2024-11-20 15:59:06.648826] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:08.513 pt2 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.513 malloc3 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.513 [2024-11-20 15:59:06.704080] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:11:08.513 [2024-11-20 15:59:06.704129] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:08.513 [2024-11-20 15:59:06.704149] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:11:08.513 [2024-11-20 15:59:06.704159] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:08.513 [2024-11-20 15:59:06.706317] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:08.513 [2024-11-20 15:59:06.706349] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:11:08.513 pt3 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.513 malloc4 00:11:08.513 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.514 [2024-11-20 15:59:06.741396] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:08.514 [2024-11-20 15:59:06.741441] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:08.514 [2024-11-20 15:59:06.741458] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:11:08.514 [2024-11-20 15:59:06.741466] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:08.514 [2024-11-20 15:59:06.743732] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:08.514 [2024-11-20 15:59:06.743759] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:08.514 pt4 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.514 [2024-11-20 15:59:06.749433] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:11:08.514 [2024-11-20 15:59:06.751424] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:08.514 [2024-11-20 15:59:06.751515] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:08.514 [2024-11-20 15:59:06.751566] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:08.514 [2024-11-20 15:59:06.751762] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:11:08.514 [2024-11-20 15:59:06.751773] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:08.514 [2024-11-20 15:59:06.752029] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:11:08.514 [2024-11-20 15:59:06.752180] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:11:08.514 [2024-11-20 15:59:06.752192] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:11:08.514 [2024-11-20 15:59:06.752328] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.514 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:08.771 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.771 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:08.771 "name": "raid_bdev1", 00:11:08.771 "uuid": "20b86efa-58b5-4f8f-bb6a-f1e347e18b49", 00:11:08.771 "strip_size_kb": 64, 00:11:08.771 "state": "online", 00:11:08.771 "raid_level": "raid0", 00:11:08.771 "superblock": true, 00:11:08.771 "num_base_bdevs": 4, 00:11:08.771 "num_base_bdevs_discovered": 4, 00:11:08.771 "num_base_bdevs_operational": 4, 00:11:08.771 "base_bdevs_list": [ 00:11:08.771 { 00:11:08.771 "name": "pt1", 00:11:08.771 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:08.771 "is_configured": true, 00:11:08.771 "data_offset": 2048, 00:11:08.771 "data_size": 63488 00:11:08.771 }, 00:11:08.771 { 00:11:08.771 "name": "pt2", 00:11:08.771 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:08.771 "is_configured": true, 00:11:08.771 "data_offset": 2048, 00:11:08.771 "data_size": 63488 00:11:08.771 }, 00:11:08.771 { 00:11:08.771 "name": "pt3", 00:11:08.771 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:08.771 "is_configured": true, 00:11:08.771 "data_offset": 2048, 00:11:08.771 "data_size": 63488 00:11:08.771 }, 00:11:08.771 { 00:11:08.771 "name": "pt4", 00:11:08.771 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:08.771 "is_configured": true, 00:11:08.771 "data_offset": 2048, 00:11:08.771 "data_size": 63488 00:11:08.771 } 00:11:08.771 ] 00:11:08.771 }' 00:11:08.772 15:59:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:08.772 15:59:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.029 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:11:09.029 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:11:09.029 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:09.029 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:09.029 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:09.029 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.030 [2024-11-20 15:59:07.081928] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:09.030 "name": "raid_bdev1", 00:11:09.030 "aliases": [ 00:11:09.030 "20b86efa-58b5-4f8f-bb6a-f1e347e18b49" 00:11:09.030 ], 00:11:09.030 "product_name": "Raid Volume", 00:11:09.030 "block_size": 512, 00:11:09.030 "num_blocks": 253952, 00:11:09.030 "uuid": "20b86efa-58b5-4f8f-bb6a-f1e347e18b49", 00:11:09.030 "assigned_rate_limits": { 00:11:09.030 "rw_ios_per_sec": 0, 00:11:09.030 "rw_mbytes_per_sec": 0, 00:11:09.030 "r_mbytes_per_sec": 0, 00:11:09.030 "w_mbytes_per_sec": 0 00:11:09.030 }, 00:11:09.030 "claimed": false, 00:11:09.030 "zoned": false, 00:11:09.030 "supported_io_types": { 00:11:09.030 "read": true, 00:11:09.030 "write": true, 00:11:09.030 "unmap": true, 00:11:09.030 "flush": true, 00:11:09.030 "reset": true, 00:11:09.030 "nvme_admin": false, 00:11:09.030 "nvme_io": false, 00:11:09.030 "nvme_io_md": false, 00:11:09.030 "write_zeroes": true, 00:11:09.030 "zcopy": false, 00:11:09.030 "get_zone_info": false, 00:11:09.030 "zone_management": false, 00:11:09.030 "zone_append": false, 00:11:09.030 "compare": false, 00:11:09.030 "compare_and_write": false, 00:11:09.030 "abort": false, 00:11:09.030 "seek_hole": false, 00:11:09.030 "seek_data": false, 00:11:09.030 "copy": false, 00:11:09.030 "nvme_iov_md": false 00:11:09.030 }, 00:11:09.030 "memory_domains": [ 00:11:09.030 { 00:11:09.030 "dma_device_id": "system", 00:11:09.030 "dma_device_type": 1 00:11:09.030 }, 00:11:09.030 { 00:11:09.030 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:09.030 "dma_device_type": 2 00:11:09.030 }, 00:11:09.030 { 00:11:09.030 "dma_device_id": "system", 00:11:09.030 "dma_device_type": 1 00:11:09.030 }, 00:11:09.030 { 00:11:09.030 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:09.030 "dma_device_type": 2 00:11:09.030 }, 00:11:09.030 { 00:11:09.030 "dma_device_id": "system", 00:11:09.030 "dma_device_type": 1 00:11:09.030 }, 00:11:09.030 { 00:11:09.030 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:09.030 "dma_device_type": 2 00:11:09.030 }, 00:11:09.030 { 00:11:09.030 "dma_device_id": "system", 00:11:09.030 "dma_device_type": 1 00:11:09.030 }, 00:11:09.030 { 00:11:09.030 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:09.030 "dma_device_type": 2 00:11:09.030 } 00:11:09.030 ], 00:11:09.030 "driver_specific": { 00:11:09.030 "raid": { 00:11:09.030 "uuid": "20b86efa-58b5-4f8f-bb6a-f1e347e18b49", 00:11:09.030 "strip_size_kb": 64, 00:11:09.030 "state": "online", 00:11:09.030 "raid_level": "raid0", 00:11:09.030 "superblock": true, 00:11:09.030 "num_base_bdevs": 4, 00:11:09.030 "num_base_bdevs_discovered": 4, 00:11:09.030 "num_base_bdevs_operational": 4, 00:11:09.030 "base_bdevs_list": [ 00:11:09.030 { 00:11:09.030 "name": "pt1", 00:11:09.030 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:09.030 "is_configured": true, 00:11:09.030 "data_offset": 2048, 00:11:09.030 "data_size": 63488 00:11:09.030 }, 00:11:09.030 { 00:11:09.030 "name": "pt2", 00:11:09.030 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:09.030 "is_configured": true, 00:11:09.030 "data_offset": 2048, 00:11:09.030 "data_size": 63488 00:11:09.030 }, 00:11:09.030 { 00:11:09.030 "name": "pt3", 00:11:09.030 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:09.030 "is_configured": true, 00:11:09.030 "data_offset": 2048, 00:11:09.030 "data_size": 63488 00:11:09.030 }, 00:11:09.030 { 00:11:09.030 "name": "pt4", 00:11:09.030 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:09.030 "is_configured": true, 00:11:09.030 "data_offset": 2048, 00:11:09.030 "data_size": 63488 00:11:09.030 } 00:11:09.030 ] 00:11:09.030 } 00:11:09.030 } 00:11:09.030 }' 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:11:09.030 pt2 00:11:09.030 pt3 00:11:09.030 pt4' 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:09.030 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:09.288 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:11:09.288 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.288 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.288 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:09.288 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.288 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:09.288 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:09.288 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:11:09.288 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.289 [2024-11-20 15:59:07.318000] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=20b86efa-58b5-4f8f-bb6a-f1e347e18b49 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 20b86efa-58b5-4f8f-bb6a-f1e347e18b49 ']' 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.289 [2024-11-20 15:59:07.337649] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:09.289 [2024-11-20 15:59:07.337682] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:09.289 [2024-11-20 15:59:07.337754] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:09.289 [2024-11-20 15:59:07.337829] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:09.289 [2024-11-20 15:59:07.337846] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.289 [2024-11-20 15:59:07.445811] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:11:09.289 [2024-11-20 15:59:07.447804] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:11:09.289 [2024-11-20 15:59:07.447863] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:11:09.289 [2024-11-20 15:59:07.447901] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:11:09.289 [2024-11-20 15:59:07.447952] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:11:09.289 [2024-11-20 15:59:07.448002] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:11:09.289 [2024-11-20 15:59:07.448027] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:11:09.289 [2024-11-20 15:59:07.448046] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:11:09.289 [2024-11-20 15:59:07.448058] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:09.289 [2024-11-20 15:59:07.448076] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:11:09.289 request: 00:11:09.289 { 00:11:09.289 "name": "raid_bdev1", 00:11:09.289 "raid_level": "raid0", 00:11:09.289 "base_bdevs": [ 00:11:09.289 "malloc1", 00:11:09.289 "malloc2", 00:11:09.289 "malloc3", 00:11:09.289 "malloc4" 00:11:09.289 ], 00:11:09.289 "strip_size_kb": 64, 00:11:09.289 "superblock": false, 00:11:09.289 "method": "bdev_raid_create", 00:11:09.289 "req_id": 1 00:11:09.289 } 00:11:09.289 Got JSON-RPC error response 00:11:09.289 response: 00:11:09.289 { 00:11:09.289 "code": -17, 00:11:09.289 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:11:09.289 } 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.289 [2024-11-20 15:59:07.489825] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:11:09.289 [2024-11-20 15:59:07.489888] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:09.289 [2024-11-20 15:59:07.489908] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:09.289 [2024-11-20 15:59:07.489918] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:09.289 [2024-11-20 15:59:07.492204] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:09.289 [2024-11-20 15:59:07.492241] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:11:09.289 [2024-11-20 15:59:07.492320] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:11:09.289 [2024-11-20 15:59:07.492377] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:11:09.289 pt1 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:09.289 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:09.290 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:09.290 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:09.290 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.290 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.290 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:09.290 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.290 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:09.290 "name": "raid_bdev1", 00:11:09.290 "uuid": "20b86efa-58b5-4f8f-bb6a-f1e347e18b49", 00:11:09.290 "strip_size_kb": 64, 00:11:09.290 "state": "configuring", 00:11:09.290 "raid_level": "raid0", 00:11:09.290 "superblock": true, 00:11:09.290 "num_base_bdevs": 4, 00:11:09.290 "num_base_bdevs_discovered": 1, 00:11:09.290 "num_base_bdevs_operational": 4, 00:11:09.290 "base_bdevs_list": [ 00:11:09.290 { 00:11:09.290 "name": "pt1", 00:11:09.290 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:09.290 "is_configured": true, 00:11:09.290 "data_offset": 2048, 00:11:09.290 "data_size": 63488 00:11:09.290 }, 00:11:09.290 { 00:11:09.290 "name": null, 00:11:09.290 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:09.290 "is_configured": false, 00:11:09.290 "data_offset": 2048, 00:11:09.290 "data_size": 63488 00:11:09.290 }, 00:11:09.290 { 00:11:09.290 "name": null, 00:11:09.290 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:09.290 "is_configured": false, 00:11:09.290 "data_offset": 2048, 00:11:09.290 "data_size": 63488 00:11:09.290 }, 00:11:09.290 { 00:11:09.290 "name": null, 00:11:09.290 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:09.290 "is_configured": false, 00:11:09.290 "data_offset": 2048, 00:11:09.290 "data_size": 63488 00:11:09.290 } 00:11:09.290 ] 00:11:09.290 }' 00:11:09.290 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:09.290 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.853 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:11:09.853 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.854 [2024-11-20 15:59:07.802045] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:09.854 [2024-11-20 15:59:07.802117] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:09.854 [2024-11-20 15:59:07.802138] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:11:09.854 [2024-11-20 15:59:07.802149] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:09.854 [2024-11-20 15:59:07.802568] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:09.854 [2024-11-20 15:59:07.802590] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:09.854 [2024-11-20 15:59:07.802680] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:11:09.854 [2024-11-20 15:59:07.802704] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:09.854 pt2 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.854 [2024-11-20 15:59:07.810044] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:09.854 "name": "raid_bdev1", 00:11:09.854 "uuid": "20b86efa-58b5-4f8f-bb6a-f1e347e18b49", 00:11:09.854 "strip_size_kb": 64, 00:11:09.854 "state": "configuring", 00:11:09.854 "raid_level": "raid0", 00:11:09.854 "superblock": true, 00:11:09.854 "num_base_bdevs": 4, 00:11:09.854 "num_base_bdevs_discovered": 1, 00:11:09.854 "num_base_bdevs_operational": 4, 00:11:09.854 "base_bdevs_list": [ 00:11:09.854 { 00:11:09.854 "name": "pt1", 00:11:09.854 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:09.854 "is_configured": true, 00:11:09.854 "data_offset": 2048, 00:11:09.854 "data_size": 63488 00:11:09.854 }, 00:11:09.854 { 00:11:09.854 "name": null, 00:11:09.854 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:09.854 "is_configured": false, 00:11:09.854 "data_offset": 0, 00:11:09.854 "data_size": 63488 00:11:09.854 }, 00:11:09.854 { 00:11:09.854 "name": null, 00:11:09.854 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:09.854 "is_configured": false, 00:11:09.854 "data_offset": 2048, 00:11:09.854 "data_size": 63488 00:11:09.854 }, 00:11:09.854 { 00:11:09.854 "name": null, 00:11:09.854 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:09.854 "is_configured": false, 00:11:09.854 "data_offset": 2048, 00:11:09.854 "data_size": 63488 00:11:09.854 } 00:11:09.854 ] 00:11:09.854 }' 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:09.854 15:59:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.113 [2024-11-20 15:59:08.134293] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:10.113 [2024-11-20 15:59:08.134346] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:10.113 [2024-11-20 15:59:08.134363] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:11:10.113 [2024-11-20 15:59:08.134371] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:10.113 [2024-11-20 15:59:08.134776] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:10.113 [2024-11-20 15:59:08.134790] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:10.113 [2024-11-20 15:59:08.134859] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:11:10.113 [2024-11-20 15:59:08.134878] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:10.113 pt2 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.113 [2024-11-20 15:59:08.142288] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:11:10.113 [2024-11-20 15:59:08.142328] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:10.113 [2024-11-20 15:59:08.142344] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:11:10.113 [2024-11-20 15:59:08.142352] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:10.113 [2024-11-20 15:59:08.142708] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:10.113 [2024-11-20 15:59:08.142727] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:11:10.113 [2024-11-20 15:59:08.142782] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:11:10.113 [2024-11-20 15:59:08.142802] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:10.113 pt3 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.113 [2024-11-20 15:59:08.150266] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:10.113 [2024-11-20 15:59:08.150299] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:10.113 [2024-11-20 15:59:08.150312] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:11:10.113 [2024-11-20 15:59:08.150320] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:10.113 [2024-11-20 15:59:08.150662] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:10.113 [2024-11-20 15:59:08.150694] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:10.113 [2024-11-20 15:59:08.150747] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:11:10.113 [2024-11-20 15:59:08.150764] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:10.113 [2024-11-20 15:59:08.150887] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:11:10.113 [2024-11-20 15:59:08.150896] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:10.113 [2024-11-20 15:59:08.151115] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:11:10.113 [2024-11-20 15:59:08.151245] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:11:10.113 [2024-11-20 15:59:08.151255] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:11:10.113 [2024-11-20 15:59:08.151369] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:10.113 pt4 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.113 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:10.113 "name": "raid_bdev1", 00:11:10.113 "uuid": "20b86efa-58b5-4f8f-bb6a-f1e347e18b49", 00:11:10.113 "strip_size_kb": 64, 00:11:10.113 "state": "online", 00:11:10.113 "raid_level": "raid0", 00:11:10.113 "superblock": true, 00:11:10.113 "num_base_bdevs": 4, 00:11:10.113 "num_base_bdevs_discovered": 4, 00:11:10.113 "num_base_bdevs_operational": 4, 00:11:10.113 "base_bdevs_list": [ 00:11:10.113 { 00:11:10.113 "name": "pt1", 00:11:10.113 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:10.113 "is_configured": true, 00:11:10.113 "data_offset": 2048, 00:11:10.113 "data_size": 63488 00:11:10.113 }, 00:11:10.113 { 00:11:10.113 "name": "pt2", 00:11:10.113 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:10.113 "is_configured": true, 00:11:10.113 "data_offset": 2048, 00:11:10.113 "data_size": 63488 00:11:10.113 }, 00:11:10.113 { 00:11:10.113 "name": "pt3", 00:11:10.113 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:10.113 "is_configured": true, 00:11:10.113 "data_offset": 2048, 00:11:10.113 "data_size": 63488 00:11:10.113 }, 00:11:10.113 { 00:11:10.113 "name": "pt4", 00:11:10.113 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:10.113 "is_configured": true, 00:11:10.113 "data_offset": 2048, 00:11:10.113 "data_size": 63488 00:11:10.113 } 00:11:10.113 ] 00:11:10.113 }' 00:11:10.114 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:10.114 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.388 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:11:10.388 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:11:10.388 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:10.388 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:10.388 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:10.388 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:10.388 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:10.388 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:10.388 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.388 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.388 [2024-11-20 15:59:08.462899] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:10.388 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.388 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:10.388 "name": "raid_bdev1", 00:11:10.388 "aliases": [ 00:11:10.388 "20b86efa-58b5-4f8f-bb6a-f1e347e18b49" 00:11:10.388 ], 00:11:10.388 "product_name": "Raid Volume", 00:11:10.388 "block_size": 512, 00:11:10.388 "num_blocks": 253952, 00:11:10.388 "uuid": "20b86efa-58b5-4f8f-bb6a-f1e347e18b49", 00:11:10.388 "assigned_rate_limits": { 00:11:10.388 "rw_ios_per_sec": 0, 00:11:10.388 "rw_mbytes_per_sec": 0, 00:11:10.388 "r_mbytes_per_sec": 0, 00:11:10.388 "w_mbytes_per_sec": 0 00:11:10.388 }, 00:11:10.388 "claimed": false, 00:11:10.388 "zoned": false, 00:11:10.388 "supported_io_types": { 00:11:10.388 "read": true, 00:11:10.388 "write": true, 00:11:10.388 "unmap": true, 00:11:10.388 "flush": true, 00:11:10.388 "reset": true, 00:11:10.388 "nvme_admin": false, 00:11:10.388 "nvme_io": false, 00:11:10.388 "nvme_io_md": false, 00:11:10.388 "write_zeroes": true, 00:11:10.388 "zcopy": false, 00:11:10.388 "get_zone_info": false, 00:11:10.388 "zone_management": false, 00:11:10.388 "zone_append": false, 00:11:10.388 "compare": false, 00:11:10.388 "compare_and_write": false, 00:11:10.388 "abort": false, 00:11:10.388 "seek_hole": false, 00:11:10.388 "seek_data": false, 00:11:10.388 "copy": false, 00:11:10.388 "nvme_iov_md": false 00:11:10.388 }, 00:11:10.388 "memory_domains": [ 00:11:10.388 { 00:11:10.388 "dma_device_id": "system", 00:11:10.388 "dma_device_type": 1 00:11:10.388 }, 00:11:10.388 { 00:11:10.388 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:10.388 "dma_device_type": 2 00:11:10.388 }, 00:11:10.388 { 00:11:10.388 "dma_device_id": "system", 00:11:10.388 "dma_device_type": 1 00:11:10.388 }, 00:11:10.388 { 00:11:10.388 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:10.388 "dma_device_type": 2 00:11:10.388 }, 00:11:10.388 { 00:11:10.388 "dma_device_id": "system", 00:11:10.388 "dma_device_type": 1 00:11:10.388 }, 00:11:10.388 { 00:11:10.388 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:10.388 "dma_device_type": 2 00:11:10.388 }, 00:11:10.388 { 00:11:10.388 "dma_device_id": "system", 00:11:10.388 "dma_device_type": 1 00:11:10.388 }, 00:11:10.388 { 00:11:10.388 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:10.388 "dma_device_type": 2 00:11:10.388 } 00:11:10.388 ], 00:11:10.388 "driver_specific": { 00:11:10.388 "raid": { 00:11:10.389 "uuid": "20b86efa-58b5-4f8f-bb6a-f1e347e18b49", 00:11:10.389 "strip_size_kb": 64, 00:11:10.389 "state": "online", 00:11:10.389 "raid_level": "raid0", 00:11:10.389 "superblock": true, 00:11:10.389 "num_base_bdevs": 4, 00:11:10.389 "num_base_bdevs_discovered": 4, 00:11:10.389 "num_base_bdevs_operational": 4, 00:11:10.389 "base_bdevs_list": [ 00:11:10.389 { 00:11:10.389 "name": "pt1", 00:11:10.389 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:10.389 "is_configured": true, 00:11:10.389 "data_offset": 2048, 00:11:10.389 "data_size": 63488 00:11:10.389 }, 00:11:10.389 { 00:11:10.389 "name": "pt2", 00:11:10.389 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:10.389 "is_configured": true, 00:11:10.389 "data_offset": 2048, 00:11:10.389 "data_size": 63488 00:11:10.389 }, 00:11:10.389 { 00:11:10.389 "name": "pt3", 00:11:10.389 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:10.389 "is_configured": true, 00:11:10.389 "data_offset": 2048, 00:11:10.389 "data_size": 63488 00:11:10.389 }, 00:11:10.389 { 00:11:10.389 "name": "pt4", 00:11:10.389 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:10.389 "is_configured": true, 00:11:10.389 "data_offset": 2048, 00:11:10.389 "data_size": 63488 00:11:10.389 } 00:11:10.389 ] 00:11:10.389 } 00:11:10.389 } 00:11:10.389 }' 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:11:10.389 pt2 00:11:10.389 pt3 00:11:10.389 pt4' 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.389 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.647 [2024-11-20 15:59:08.707056] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 20b86efa-58b5-4f8f-bb6a-f1e347e18b49 '!=' 20b86efa-58b5-4f8f-bb6a-f1e347e18b49 ']' 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 68838 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 68838 ']' 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 68838 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 68838 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:10.647 killing process with pid 68838 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 68838' 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 68838 00:11:10.647 [2024-11-20 15:59:08.762075] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:10.647 [2024-11-20 15:59:08.762151] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:10.647 [2024-11-20 15:59:08.762225] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:10.647 [2024-11-20 15:59:08.762235] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:11:10.647 15:59:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 68838 00:11:10.905 [2024-11-20 15:59:09.004905] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:11.470 ************************************ 00:11:11.470 END TEST raid_superblock_test 00:11:11.470 ************************************ 00:11:11.470 15:59:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:11:11.470 00:11:11.470 real 0m4.042s 00:11:11.470 user 0m5.784s 00:11:11.470 sys 0m0.637s 00:11:11.470 15:59:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:11.470 15:59:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:11.727 15:59:09 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 4 read 00:11:11.727 15:59:09 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:11.727 15:59:09 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:11.727 15:59:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:11.727 ************************************ 00:11:11.727 START TEST raid_read_error_test 00:11:11.727 ************************************ 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 4 read 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.RhMbKnEmNd 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=69086 00:11:11.727 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 69086 00:11:11.728 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:11.728 15:59:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 69086 ']' 00:11:11.728 15:59:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:11.728 15:59:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:11.728 15:59:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:11.728 15:59:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:11.728 15:59:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:11.728 15:59:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:11:11.728 [2024-11-20 15:59:09.824639] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:11:11.728 [2024-11-20 15:59:09.824771] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69086 ] 00:11:11.984 [2024-11-20 15:59:09.981999] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:11.984 [2024-11-20 15:59:10.081519] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:11.984 [2024-11-20 15:59:10.216695] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:11.984 [2024-11-20 15:59:10.216904] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.554 BaseBdev1_malloc 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.554 true 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.554 [2024-11-20 15:59:10.738197] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:12.554 [2024-11-20 15:59:10.738350] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:12.554 [2024-11-20 15:59:10.738375] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:11:12.554 [2024-11-20 15:59:10.738386] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:12.554 [2024-11-20 15:59:10.740514] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:12.554 BaseBdev1 00:11:12.554 [2024-11-20 15:59:10.740646] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.554 BaseBdev2_malloc 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.554 true 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.554 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.554 [2024-11-20 15:59:10.782336] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:12.554 [2024-11-20 15:59:10.782475] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:12.554 [2024-11-20 15:59:10.782498] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:12.554 [2024-11-20 15:59:10.782510] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:12.554 [2024-11-20 15:59:10.784598] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:12.555 [2024-11-20 15:59:10.784714] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:12.555 BaseBdev2 00:11:12.555 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.555 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:12.555 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:12.555 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.555 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.818 BaseBdev3_malloc 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.818 true 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.818 [2024-11-20 15:59:10.838288] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:11:12.818 [2024-11-20 15:59:10.838432] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:12.818 [2024-11-20 15:59:10.838472] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:11:12.818 [2024-11-20 15:59:10.838880] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:12.818 [2024-11-20 15:59:10.845480] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:12.818 [2024-11-20 15:59:10.845831] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:12.818 BaseBdev3 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.818 BaseBdev4_malloc 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.818 true 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.818 [2024-11-20 15:59:10.887312] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:11:12.818 [2024-11-20 15:59:10.887361] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:12.818 [2024-11-20 15:59:10.887379] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:12.818 [2024-11-20 15:59:10.887389] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:12.818 [2024-11-20 15:59:10.889517] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:12.818 [2024-11-20 15:59:10.889642] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:12.818 BaseBdev4 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.818 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.818 [2024-11-20 15:59:10.895381] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:12.818 [2024-11-20 15:59:10.897323] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:12.818 [2024-11-20 15:59:10.897481] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:12.818 [2024-11-20 15:59:10.897574] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:12.818 [2024-11-20 15:59:10.897934] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:11:12.818 [2024-11-20 15:59:10.897956] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:12.819 [2024-11-20 15:59:10.898218] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:11:12.819 [2024-11-20 15:59:10.898364] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:11:12.819 [2024-11-20 15:59:10.898375] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:11:12.819 [2024-11-20 15:59:10.898527] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:12.819 "name": "raid_bdev1", 00:11:12.819 "uuid": "f7bc9f16-989c-4425-8975-7e3e15226bbc", 00:11:12.819 "strip_size_kb": 64, 00:11:12.819 "state": "online", 00:11:12.819 "raid_level": "raid0", 00:11:12.819 "superblock": true, 00:11:12.819 "num_base_bdevs": 4, 00:11:12.819 "num_base_bdevs_discovered": 4, 00:11:12.819 "num_base_bdevs_operational": 4, 00:11:12.819 "base_bdevs_list": [ 00:11:12.819 { 00:11:12.819 "name": "BaseBdev1", 00:11:12.819 "uuid": "218b2180-a276-59ef-abfb-944536190079", 00:11:12.819 "is_configured": true, 00:11:12.819 "data_offset": 2048, 00:11:12.819 "data_size": 63488 00:11:12.819 }, 00:11:12.819 { 00:11:12.819 "name": "BaseBdev2", 00:11:12.819 "uuid": "0efc7c93-1a66-551a-8cd6-b022230685e3", 00:11:12.819 "is_configured": true, 00:11:12.819 "data_offset": 2048, 00:11:12.819 "data_size": 63488 00:11:12.819 }, 00:11:12.819 { 00:11:12.819 "name": "BaseBdev3", 00:11:12.819 "uuid": "f6f9d564-9c4b-515c-8e5c-cd6329dec99f", 00:11:12.819 "is_configured": true, 00:11:12.819 "data_offset": 2048, 00:11:12.819 "data_size": 63488 00:11:12.819 }, 00:11:12.819 { 00:11:12.819 "name": "BaseBdev4", 00:11:12.819 "uuid": "f3c4b4f3-b6d4-545e-8758-f4cd92b40b8d", 00:11:12.819 "is_configured": true, 00:11:12.819 "data_offset": 2048, 00:11:12.819 "data_size": 63488 00:11:12.819 } 00:11:12.819 ] 00:11:12.819 }' 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:12.819 15:59:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.122 15:59:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:11:13.122 15:59:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:13.122 [2024-11-20 15:59:11.288387] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006a40 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:14.062 "name": "raid_bdev1", 00:11:14.062 "uuid": "f7bc9f16-989c-4425-8975-7e3e15226bbc", 00:11:14.062 "strip_size_kb": 64, 00:11:14.062 "state": "online", 00:11:14.062 "raid_level": "raid0", 00:11:14.062 "superblock": true, 00:11:14.062 "num_base_bdevs": 4, 00:11:14.062 "num_base_bdevs_discovered": 4, 00:11:14.062 "num_base_bdevs_operational": 4, 00:11:14.062 "base_bdevs_list": [ 00:11:14.062 { 00:11:14.062 "name": "BaseBdev1", 00:11:14.062 "uuid": "218b2180-a276-59ef-abfb-944536190079", 00:11:14.062 "is_configured": true, 00:11:14.062 "data_offset": 2048, 00:11:14.062 "data_size": 63488 00:11:14.062 }, 00:11:14.062 { 00:11:14.062 "name": "BaseBdev2", 00:11:14.062 "uuid": "0efc7c93-1a66-551a-8cd6-b022230685e3", 00:11:14.062 "is_configured": true, 00:11:14.062 "data_offset": 2048, 00:11:14.062 "data_size": 63488 00:11:14.062 }, 00:11:14.062 { 00:11:14.062 "name": "BaseBdev3", 00:11:14.062 "uuid": "f6f9d564-9c4b-515c-8e5c-cd6329dec99f", 00:11:14.062 "is_configured": true, 00:11:14.062 "data_offset": 2048, 00:11:14.062 "data_size": 63488 00:11:14.062 }, 00:11:14.062 { 00:11:14.062 "name": "BaseBdev4", 00:11:14.062 "uuid": "f3c4b4f3-b6d4-545e-8758-f4cd92b40b8d", 00:11:14.062 "is_configured": true, 00:11:14.062 "data_offset": 2048, 00:11:14.062 "data_size": 63488 00:11:14.062 } 00:11:14.062 ] 00:11:14.062 }' 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:14.062 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:14.319 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:14.319 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:14.319 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:14.319 [2024-11-20 15:59:12.522453] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:14.320 [2024-11-20 15:59:12.522483] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:14.320 [2024-11-20 15:59:12.525584] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:14.320 [2024-11-20 15:59:12.525644] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:14.320 [2024-11-20 15:59:12.525698] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:14.320 [2024-11-20 15:59:12.525710] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:11:14.320 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:14.320 { 00:11:14.320 "results": [ 00:11:14.320 { 00:11:14.320 "job": "raid_bdev1", 00:11:14.320 "core_mask": "0x1", 00:11:14.320 "workload": "randrw", 00:11:14.320 "percentage": 50, 00:11:14.320 "status": "finished", 00:11:14.320 "queue_depth": 1, 00:11:14.320 "io_size": 131072, 00:11:14.320 "runtime": 1.232094, 00:11:14.320 "iops": 14642.55162349626, 00:11:14.320 "mibps": 1830.3189529370325, 00:11:14.320 "io_failed": 1, 00:11:14.320 "io_timeout": 0, 00:11:14.320 "avg_latency_us": 93.1062611172222, 00:11:14.320 "min_latency_us": 33.47692307692308, 00:11:14.320 "max_latency_us": 1688.8123076923077 00:11:14.320 } 00:11:14.320 ], 00:11:14.320 "core_count": 1 00:11:14.320 } 00:11:14.320 15:59:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 69086 00:11:14.320 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 69086 ']' 00:11:14.320 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 69086 00:11:14.320 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:11:14.320 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:14.320 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69086 00:11:14.320 killing process with pid 69086 00:11:14.320 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:14.320 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:14.320 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69086' 00:11:14.320 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 69086 00:11:14.320 [2024-11-20 15:59:12.560604] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:14.320 15:59:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 69086 00:11:14.577 [2024-11-20 15:59:12.763750] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:15.613 15:59:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:11:15.613 15:59:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.RhMbKnEmNd 00:11:15.613 15:59:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:11:15.613 15:59:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.81 00:11:15.613 15:59:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:11:15.613 15:59:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:15.613 15:59:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:11:15.613 15:59:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.81 != \0\.\0\0 ]] 00:11:15.613 00:11:15.613 real 0m3.777s 00:11:15.613 user 0m4.478s 00:11:15.613 sys 0m0.385s 00:11:15.613 ************************************ 00:11:15.613 END TEST raid_read_error_test 00:11:15.613 ************************************ 00:11:15.613 15:59:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:15.613 15:59:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:15.613 15:59:13 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 4 write 00:11:15.613 15:59:13 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:15.613 15:59:13 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:15.613 15:59:13 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:15.613 ************************************ 00:11:15.613 START TEST raid_write_error_test 00:11:15.613 ************************************ 00:11:15.613 15:59:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 4 write 00:11:15.613 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:11:15.613 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:11:15.613 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:11:15.613 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:11:15.613 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:15.613 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:11:15.613 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:15.613 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:15.613 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:11:15.613 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.xS10dRHBk8 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=69222 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 69222 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 69222 ']' 00:11:15.614 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:15.614 15:59:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:15.614 [2024-11-20 15:59:13.668365] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:11:15.614 [2024-11-20 15:59:13.668478] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69222 ] 00:11:15.614 [2024-11-20 15:59:13.826448] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:15.871 [2024-11-20 15:59:13.928472] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:15.871 [2024-11-20 15:59:14.063950] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:15.871 [2024-11-20 15:59:14.063998] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.440 BaseBdev1_malloc 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.440 true 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.440 [2024-11-20 15:59:14.557873] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:16.440 [2024-11-20 15:59:14.558033] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:16.440 [2024-11-20 15:59:14.558059] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:11:16.440 [2024-11-20 15:59:14.558072] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:16.440 [2024-11-20 15:59:14.560225] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:16.440 [2024-11-20 15:59:14.560261] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:16.440 BaseBdev1 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.440 BaseBdev2_malloc 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.440 true 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.440 [2024-11-20 15:59:14.601875] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:16.440 [2024-11-20 15:59:14.601922] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:16.440 [2024-11-20 15:59:14.601937] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:16.440 [2024-11-20 15:59:14.601946] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:16.440 [2024-11-20 15:59:14.604051] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:16.440 [2024-11-20 15:59:14.604086] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:16.440 BaseBdev2 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.440 BaseBdev3_malloc 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.440 true 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.440 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.441 [2024-11-20 15:59:14.654640] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:11:16.441 [2024-11-20 15:59:14.654704] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:16.441 [2024-11-20 15:59:14.654723] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:11:16.441 [2024-11-20 15:59:14.654734] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:16.441 [2024-11-20 15:59:14.656906] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:16.441 [2024-11-20 15:59:14.657043] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:16.441 BaseBdev3 00:11:16.441 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.441 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:16.441 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:16.441 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.441 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.441 BaseBdev4_malloc 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.698 true 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.698 [2024-11-20 15:59:14.702821] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:11:16.698 [2024-11-20 15:59:14.702863] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:16.698 [2024-11-20 15:59:14.702880] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:16.698 [2024-11-20 15:59:14.702891] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:16.698 [2024-11-20 15:59:14.705126] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:16.698 [2024-11-20 15:59:14.705167] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:16.698 BaseBdev4 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.698 [2024-11-20 15:59:14.710886] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:16.698 [2024-11-20 15:59:14.712756] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:16.698 [2024-11-20 15:59:14.712832] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:16.698 [2024-11-20 15:59:14.712897] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:16.698 [2024-11-20 15:59:14.713111] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:11:16.698 [2024-11-20 15:59:14.713131] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:16.698 [2024-11-20 15:59:14.713384] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:11:16.698 [2024-11-20 15:59:14.713535] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:11:16.698 [2024-11-20 15:59:14.713546] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:11:16.698 [2024-11-20 15:59:14.713706] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.698 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:16.698 "name": "raid_bdev1", 00:11:16.698 "uuid": "ebf95535-b497-48c5-bbd7-34023fa36be7", 00:11:16.698 "strip_size_kb": 64, 00:11:16.698 "state": "online", 00:11:16.698 "raid_level": "raid0", 00:11:16.698 "superblock": true, 00:11:16.698 "num_base_bdevs": 4, 00:11:16.698 "num_base_bdevs_discovered": 4, 00:11:16.698 "num_base_bdevs_operational": 4, 00:11:16.698 "base_bdevs_list": [ 00:11:16.698 { 00:11:16.698 "name": "BaseBdev1", 00:11:16.698 "uuid": "692ee15e-8dd3-5c26-b2e5-e069bb5bebda", 00:11:16.698 "is_configured": true, 00:11:16.698 "data_offset": 2048, 00:11:16.698 "data_size": 63488 00:11:16.698 }, 00:11:16.698 { 00:11:16.698 "name": "BaseBdev2", 00:11:16.699 "uuid": "cdb24868-b453-5c4f-baa4-f534399df694", 00:11:16.699 "is_configured": true, 00:11:16.699 "data_offset": 2048, 00:11:16.699 "data_size": 63488 00:11:16.699 }, 00:11:16.699 { 00:11:16.699 "name": "BaseBdev3", 00:11:16.699 "uuid": "e7ff882c-1a19-540f-a05a-82c6dba61f0a", 00:11:16.699 "is_configured": true, 00:11:16.699 "data_offset": 2048, 00:11:16.699 "data_size": 63488 00:11:16.699 }, 00:11:16.699 { 00:11:16.699 "name": "BaseBdev4", 00:11:16.699 "uuid": "1943f80f-8e47-5b5e-88ae-dfad8a493abe", 00:11:16.699 "is_configured": true, 00:11:16.699 "data_offset": 2048, 00:11:16.699 "data_size": 63488 00:11:16.699 } 00:11:16.699 ] 00:11:16.699 }' 00:11:16.699 15:59:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:16.699 15:59:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.959 15:59:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:11:16.959 15:59:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:16.959 [2024-11-20 15:59:15.111923] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006a40 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.895 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:17.895 "name": "raid_bdev1", 00:11:17.896 "uuid": "ebf95535-b497-48c5-bbd7-34023fa36be7", 00:11:17.896 "strip_size_kb": 64, 00:11:17.896 "state": "online", 00:11:17.896 "raid_level": "raid0", 00:11:17.896 "superblock": true, 00:11:17.896 "num_base_bdevs": 4, 00:11:17.896 "num_base_bdevs_discovered": 4, 00:11:17.896 "num_base_bdevs_operational": 4, 00:11:17.896 "base_bdevs_list": [ 00:11:17.896 { 00:11:17.896 "name": "BaseBdev1", 00:11:17.896 "uuid": "692ee15e-8dd3-5c26-b2e5-e069bb5bebda", 00:11:17.896 "is_configured": true, 00:11:17.896 "data_offset": 2048, 00:11:17.896 "data_size": 63488 00:11:17.896 }, 00:11:17.896 { 00:11:17.896 "name": "BaseBdev2", 00:11:17.896 "uuid": "cdb24868-b453-5c4f-baa4-f534399df694", 00:11:17.896 "is_configured": true, 00:11:17.896 "data_offset": 2048, 00:11:17.896 "data_size": 63488 00:11:17.896 }, 00:11:17.896 { 00:11:17.896 "name": "BaseBdev3", 00:11:17.896 "uuid": "e7ff882c-1a19-540f-a05a-82c6dba61f0a", 00:11:17.896 "is_configured": true, 00:11:17.896 "data_offset": 2048, 00:11:17.896 "data_size": 63488 00:11:17.896 }, 00:11:17.896 { 00:11:17.896 "name": "BaseBdev4", 00:11:17.896 "uuid": "1943f80f-8e47-5b5e-88ae-dfad8a493abe", 00:11:17.896 "is_configured": true, 00:11:17.896 "data_offset": 2048, 00:11:17.896 "data_size": 63488 00:11:17.896 } 00:11:17.896 ] 00:11:17.896 }' 00:11:17.896 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:17.896 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:18.154 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:18.154 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:18.154 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:18.154 [2024-11-20 15:59:16.374110] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:18.154 [2024-11-20 15:59:16.374142] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:18.154 [2024-11-20 15:59:16.377220] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:18.154 [2024-11-20 15:59:16.377281] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:18.154 [2024-11-20 15:59:16.377325] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:18.154 [2024-11-20 15:59:16.377336] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:11:18.154 { 00:11:18.154 "results": [ 00:11:18.154 { 00:11:18.154 "job": "raid_bdev1", 00:11:18.154 "core_mask": "0x1", 00:11:18.154 "workload": "randrw", 00:11:18.154 "percentage": 50, 00:11:18.154 "status": "finished", 00:11:18.154 "queue_depth": 1, 00:11:18.154 "io_size": 131072, 00:11:18.154 "runtime": 1.260361, 00:11:18.154 "iops": 14303.838344728218, 00:11:18.154 "mibps": 1787.9797930910272, 00:11:18.154 "io_failed": 1, 00:11:18.154 "io_timeout": 0, 00:11:18.154 "avg_latency_us": 95.49894878763702, 00:11:18.154 "min_latency_us": 33.673846153846156, 00:11:18.154 "max_latency_us": 1688.8123076923077 00:11:18.154 } 00:11:18.154 ], 00:11:18.154 "core_count": 1 00:11:18.154 } 00:11:18.154 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:18.154 15:59:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 69222 00:11:18.154 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 69222 ']' 00:11:18.154 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 69222 00:11:18.154 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:11:18.154 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:18.154 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69222 00:11:18.420 killing process with pid 69222 00:11:18.420 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:18.420 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:18.420 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69222' 00:11:18.420 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 69222 00:11:18.420 [2024-11-20 15:59:16.409217] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:18.420 15:59:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 69222 00:11:18.420 [2024-11-20 15:59:16.611662] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:19.359 15:59:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.xS10dRHBk8 00:11:19.359 15:59:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:11:19.360 15:59:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:11:19.360 15:59:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.79 00:11:19.360 15:59:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:11:19.360 15:59:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:19.360 15:59:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:11:19.360 15:59:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.79 != \0\.\0\0 ]] 00:11:19.360 00:11:19.360 real 0m3.786s 00:11:19.360 user 0m4.456s 00:11:19.360 sys 0m0.414s 00:11:19.360 15:59:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:19.360 15:59:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.360 ************************************ 00:11:19.360 END TEST raid_write_error_test 00:11:19.360 ************************************ 00:11:19.360 15:59:17 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:11:19.360 15:59:17 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 4 false 00:11:19.360 15:59:17 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:19.360 15:59:17 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:19.360 15:59:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:19.360 ************************************ 00:11:19.360 START TEST raid_state_function_test 00:11:19.360 ************************************ 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 4 false 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=69353 00:11:19.360 Process raid pid: 69353 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 69353' 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 69353 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 69353 ']' 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:19.360 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:19.360 15:59:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.360 [2024-11-20 15:59:17.489708] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:11:19.360 [2024-11-20 15:59:17.489822] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:19.621 [2024-11-20 15:59:17.648297] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:19.621 [2024-11-20 15:59:17.750853] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:19.881 [2024-11-20 15:59:17.888241] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:19.881 [2024-11-20 15:59:17.888275] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.143 [2024-11-20 15:59:18.343288] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:20.143 [2024-11-20 15:59:18.343338] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:20.143 [2024-11-20 15:59:18.343349] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:20.143 [2024-11-20 15:59:18.343359] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:20.143 [2024-11-20 15:59:18.343366] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:20.143 [2024-11-20 15:59:18.343376] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:20.143 [2024-11-20 15:59:18.343383] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:20.143 [2024-11-20 15:59:18.343392] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.143 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:20.143 "name": "Existed_Raid", 00:11:20.143 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:20.143 "strip_size_kb": 64, 00:11:20.143 "state": "configuring", 00:11:20.143 "raid_level": "concat", 00:11:20.143 "superblock": false, 00:11:20.143 "num_base_bdevs": 4, 00:11:20.143 "num_base_bdevs_discovered": 0, 00:11:20.143 "num_base_bdevs_operational": 4, 00:11:20.143 "base_bdevs_list": [ 00:11:20.143 { 00:11:20.143 "name": "BaseBdev1", 00:11:20.143 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:20.143 "is_configured": false, 00:11:20.143 "data_offset": 0, 00:11:20.143 "data_size": 0 00:11:20.143 }, 00:11:20.143 { 00:11:20.143 "name": "BaseBdev2", 00:11:20.143 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:20.143 "is_configured": false, 00:11:20.143 "data_offset": 0, 00:11:20.143 "data_size": 0 00:11:20.143 }, 00:11:20.143 { 00:11:20.143 "name": "BaseBdev3", 00:11:20.143 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:20.143 "is_configured": false, 00:11:20.144 "data_offset": 0, 00:11:20.144 "data_size": 0 00:11:20.144 }, 00:11:20.144 { 00:11:20.144 "name": "BaseBdev4", 00:11:20.144 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:20.144 "is_configured": false, 00:11:20.144 "data_offset": 0, 00:11:20.144 "data_size": 0 00:11:20.144 } 00:11:20.144 ] 00:11:20.144 }' 00:11:20.144 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:20.144 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.778 [2024-11-20 15:59:18.675298] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:20.778 [2024-11-20 15:59:18.675335] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.778 [2024-11-20 15:59:18.683307] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:20.778 [2024-11-20 15:59:18.683345] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:20.778 [2024-11-20 15:59:18.683354] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:20.778 [2024-11-20 15:59:18.683362] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:20.778 [2024-11-20 15:59:18.683369] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:20.778 [2024-11-20 15:59:18.683377] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:20.778 [2024-11-20 15:59:18.683383] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:20.778 [2024-11-20 15:59:18.683391] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.778 [2024-11-20 15:59:18.719610] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:20.778 BaseBdev1 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.778 [ 00:11:20.778 { 00:11:20.778 "name": "BaseBdev1", 00:11:20.778 "aliases": [ 00:11:20.778 "c4dfb73c-2cae-4b09-ade8-2e61bf9143e0" 00:11:20.778 ], 00:11:20.778 "product_name": "Malloc disk", 00:11:20.778 "block_size": 512, 00:11:20.778 "num_blocks": 65536, 00:11:20.778 "uuid": "c4dfb73c-2cae-4b09-ade8-2e61bf9143e0", 00:11:20.778 "assigned_rate_limits": { 00:11:20.778 "rw_ios_per_sec": 0, 00:11:20.778 "rw_mbytes_per_sec": 0, 00:11:20.778 "r_mbytes_per_sec": 0, 00:11:20.778 "w_mbytes_per_sec": 0 00:11:20.778 }, 00:11:20.778 "claimed": true, 00:11:20.778 "claim_type": "exclusive_write", 00:11:20.778 "zoned": false, 00:11:20.778 "supported_io_types": { 00:11:20.778 "read": true, 00:11:20.778 "write": true, 00:11:20.778 "unmap": true, 00:11:20.778 "flush": true, 00:11:20.778 "reset": true, 00:11:20.778 "nvme_admin": false, 00:11:20.778 "nvme_io": false, 00:11:20.778 "nvme_io_md": false, 00:11:20.778 "write_zeroes": true, 00:11:20.778 "zcopy": true, 00:11:20.778 "get_zone_info": false, 00:11:20.778 "zone_management": false, 00:11:20.778 "zone_append": false, 00:11:20.778 "compare": false, 00:11:20.778 "compare_and_write": false, 00:11:20.778 "abort": true, 00:11:20.778 "seek_hole": false, 00:11:20.778 "seek_data": false, 00:11:20.778 "copy": true, 00:11:20.778 "nvme_iov_md": false 00:11:20.778 }, 00:11:20.778 "memory_domains": [ 00:11:20.778 { 00:11:20.778 "dma_device_id": "system", 00:11:20.778 "dma_device_type": 1 00:11:20.778 }, 00:11:20.778 { 00:11:20.778 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:20.778 "dma_device_type": 2 00:11:20.778 } 00:11:20.778 ], 00:11:20.778 "driver_specific": {} 00:11:20.778 } 00:11:20.778 ] 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.778 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:20.778 "name": "Existed_Raid", 00:11:20.778 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:20.779 "strip_size_kb": 64, 00:11:20.779 "state": "configuring", 00:11:20.779 "raid_level": "concat", 00:11:20.779 "superblock": false, 00:11:20.779 "num_base_bdevs": 4, 00:11:20.779 "num_base_bdevs_discovered": 1, 00:11:20.779 "num_base_bdevs_operational": 4, 00:11:20.779 "base_bdevs_list": [ 00:11:20.779 { 00:11:20.779 "name": "BaseBdev1", 00:11:20.779 "uuid": "c4dfb73c-2cae-4b09-ade8-2e61bf9143e0", 00:11:20.779 "is_configured": true, 00:11:20.779 "data_offset": 0, 00:11:20.779 "data_size": 65536 00:11:20.779 }, 00:11:20.779 { 00:11:20.779 "name": "BaseBdev2", 00:11:20.779 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:20.779 "is_configured": false, 00:11:20.779 "data_offset": 0, 00:11:20.779 "data_size": 0 00:11:20.779 }, 00:11:20.779 { 00:11:20.779 "name": "BaseBdev3", 00:11:20.779 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:20.779 "is_configured": false, 00:11:20.779 "data_offset": 0, 00:11:20.779 "data_size": 0 00:11:20.779 }, 00:11:20.779 { 00:11:20.779 "name": "BaseBdev4", 00:11:20.779 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:20.779 "is_configured": false, 00:11:20.779 "data_offset": 0, 00:11:20.779 "data_size": 0 00:11:20.779 } 00:11:20.779 ] 00:11:20.779 }' 00:11:20.779 15:59:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:20.779 15:59:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.059 [2024-11-20 15:59:19.095733] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:21.059 [2024-11-20 15:59:19.095781] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.059 [2024-11-20 15:59:19.107803] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:21.059 [2024-11-20 15:59:19.109636] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:21.059 [2024-11-20 15:59:19.109691] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:21.059 [2024-11-20 15:59:19.109700] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:21.059 [2024-11-20 15:59:19.109712] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:21.059 [2024-11-20 15:59:19.109718] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:21.059 [2024-11-20 15:59:19.109726] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:21.059 "name": "Existed_Raid", 00:11:21.059 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.059 "strip_size_kb": 64, 00:11:21.059 "state": "configuring", 00:11:21.059 "raid_level": "concat", 00:11:21.059 "superblock": false, 00:11:21.059 "num_base_bdevs": 4, 00:11:21.059 "num_base_bdevs_discovered": 1, 00:11:21.059 "num_base_bdevs_operational": 4, 00:11:21.059 "base_bdevs_list": [ 00:11:21.059 { 00:11:21.059 "name": "BaseBdev1", 00:11:21.059 "uuid": "c4dfb73c-2cae-4b09-ade8-2e61bf9143e0", 00:11:21.059 "is_configured": true, 00:11:21.059 "data_offset": 0, 00:11:21.059 "data_size": 65536 00:11:21.059 }, 00:11:21.059 { 00:11:21.059 "name": "BaseBdev2", 00:11:21.059 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.059 "is_configured": false, 00:11:21.059 "data_offset": 0, 00:11:21.059 "data_size": 0 00:11:21.059 }, 00:11:21.059 { 00:11:21.059 "name": "BaseBdev3", 00:11:21.059 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.059 "is_configured": false, 00:11:21.059 "data_offset": 0, 00:11:21.059 "data_size": 0 00:11:21.059 }, 00:11:21.059 { 00:11:21.059 "name": "BaseBdev4", 00:11:21.059 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.059 "is_configured": false, 00:11:21.059 "data_offset": 0, 00:11:21.059 "data_size": 0 00:11:21.059 } 00:11:21.059 ] 00:11:21.059 }' 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:21.059 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.320 [2024-11-20 15:59:19.454400] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:21.320 BaseBdev2 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.320 [ 00:11:21.320 { 00:11:21.320 "name": "BaseBdev2", 00:11:21.320 "aliases": [ 00:11:21.320 "f6ed7921-d6cf-4973-b8f9-59c67a038d1f" 00:11:21.320 ], 00:11:21.320 "product_name": "Malloc disk", 00:11:21.320 "block_size": 512, 00:11:21.320 "num_blocks": 65536, 00:11:21.320 "uuid": "f6ed7921-d6cf-4973-b8f9-59c67a038d1f", 00:11:21.320 "assigned_rate_limits": { 00:11:21.320 "rw_ios_per_sec": 0, 00:11:21.320 "rw_mbytes_per_sec": 0, 00:11:21.320 "r_mbytes_per_sec": 0, 00:11:21.320 "w_mbytes_per_sec": 0 00:11:21.320 }, 00:11:21.320 "claimed": true, 00:11:21.320 "claim_type": "exclusive_write", 00:11:21.320 "zoned": false, 00:11:21.320 "supported_io_types": { 00:11:21.320 "read": true, 00:11:21.320 "write": true, 00:11:21.320 "unmap": true, 00:11:21.320 "flush": true, 00:11:21.320 "reset": true, 00:11:21.320 "nvme_admin": false, 00:11:21.320 "nvme_io": false, 00:11:21.320 "nvme_io_md": false, 00:11:21.320 "write_zeroes": true, 00:11:21.320 "zcopy": true, 00:11:21.320 "get_zone_info": false, 00:11:21.320 "zone_management": false, 00:11:21.320 "zone_append": false, 00:11:21.320 "compare": false, 00:11:21.320 "compare_and_write": false, 00:11:21.320 "abort": true, 00:11:21.320 "seek_hole": false, 00:11:21.320 "seek_data": false, 00:11:21.320 "copy": true, 00:11:21.320 "nvme_iov_md": false 00:11:21.320 }, 00:11:21.320 "memory_domains": [ 00:11:21.320 { 00:11:21.320 "dma_device_id": "system", 00:11:21.320 "dma_device_type": 1 00:11:21.320 }, 00:11:21.320 { 00:11:21.320 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:21.320 "dma_device_type": 2 00:11:21.320 } 00:11:21.320 ], 00:11:21.320 "driver_specific": {} 00:11:21.320 } 00:11:21.320 ] 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:21.320 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:21.321 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:21.321 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:21.321 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:21.321 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:21.321 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.321 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.321 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.321 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:21.321 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.321 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:21.321 "name": "Existed_Raid", 00:11:21.321 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.321 "strip_size_kb": 64, 00:11:21.321 "state": "configuring", 00:11:21.321 "raid_level": "concat", 00:11:21.321 "superblock": false, 00:11:21.321 "num_base_bdevs": 4, 00:11:21.321 "num_base_bdevs_discovered": 2, 00:11:21.321 "num_base_bdevs_operational": 4, 00:11:21.321 "base_bdevs_list": [ 00:11:21.321 { 00:11:21.321 "name": "BaseBdev1", 00:11:21.321 "uuid": "c4dfb73c-2cae-4b09-ade8-2e61bf9143e0", 00:11:21.321 "is_configured": true, 00:11:21.321 "data_offset": 0, 00:11:21.321 "data_size": 65536 00:11:21.321 }, 00:11:21.321 { 00:11:21.321 "name": "BaseBdev2", 00:11:21.321 "uuid": "f6ed7921-d6cf-4973-b8f9-59c67a038d1f", 00:11:21.321 "is_configured": true, 00:11:21.321 "data_offset": 0, 00:11:21.321 "data_size": 65536 00:11:21.321 }, 00:11:21.321 { 00:11:21.321 "name": "BaseBdev3", 00:11:21.321 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.321 "is_configured": false, 00:11:21.321 "data_offset": 0, 00:11:21.321 "data_size": 0 00:11:21.321 }, 00:11:21.321 { 00:11:21.321 "name": "BaseBdev4", 00:11:21.321 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.321 "is_configured": false, 00:11:21.321 "data_offset": 0, 00:11:21.321 "data_size": 0 00:11:21.321 } 00:11:21.321 ] 00:11:21.321 }' 00:11:21.321 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:21.321 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.581 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:21.581 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.581 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.841 [2024-11-20 15:59:19.840209] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:21.841 BaseBdev3 00:11:21.841 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.841 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:11:21.841 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:11:21.841 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:21.841 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:21.841 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:21.841 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:21.841 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:21.841 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.841 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.841 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.841 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:21.841 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.841 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.841 [ 00:11:21.841 { 00:11:21.841 "name": "BaseBdev3", 00:11:21.841 "aliases": [ 00:11:21.841 "4f06b301-44df-464c-801f-aa1c09193a96" 00:11:21.841 ], 00:11:21.841 "product_name": "Malloc disk", 00:11:21.841 "block_size": 512, 00:11:21.841 "num_blocks": 65536, 00:11:21.841 "uuid": "4f06b301-44df-464c-801f-aa1c09193a96", 00:11:21.841 "assigned_rate_limits": { 00:11:21.841 "rw_ios_per_sec": 0, 00:11:21.841 "rw_mbytes_per_sec": 0, 00:11:21.841 "r_mbytes_per_sec": 0, 00:11:21.841 "w_mbytes_per_sec": 0 00:11:21.841 }, 00:11:21.841 "claimed": true, 00:11:21.841 "claim_type": "exclusive_write", 00:11:21.841 "zoned": false, 00:11:21.841 "supported_io_types": { 00:11:21.841 "read": true, 00:11:21.841 "write": true, 00:11:21.841 "unmap": true, 00:11:21.841 "flush": true, 00:11:21.841 "reset": true, 00:11:21.841 "nvme_admin": false, 00:11:21.841 "nvme_io": false, 00:11:21.841 "nvme_io_md": false, 00:11:21.841 "write_zeroes": true, 00:11:21.841 "zcopy": true, 00:11:21.841 "get_zone_info": false, 00:11:21.841 "zone_management": false, 00:11:21.841 "zone_append": false, 00:11:21.841 "compare": false, 00:11:21.841 "compare_and_write": false, 00:11:21.841 "abort": true, 00:11:21.841 "seek_hole": false, 00:11:21.841 "seek_data": false, 00:11:21.841 "copy": true, 00:11:21.841 "nvme_iov_md": false 00:11:21.841 }, 00:11:21.841 "memory_domains": [ 00:11:21.841 { 00:11:21.841 "dma_device_id": "system", 00:11:21.841 "dma_device_type": 1 00:11:21.841 }, 00:11:21.841 { 00:11:21.841 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:21.841 "dma_device_type": 2 00:11:21.841 } 00:11:21.841 ], 00:11:21.841 "driver_specific": {} 00:11:21.841 } 00:11:21.841 ] 00:11:21.841 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:21.842 "name": "Existed_Raid", 00:11:21.842 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.842 "strip_size_kb": 64, 00:11:21.842 "state": "configuring", 00:11:21.842 "raid_level": "concat", 00:11:21.842 "superblock": false, 00:11:21.842 "num_base_bdevs": 4, 00:11:21.842 "num_base_bdevs_discovered": 3, 00:11:21.842 "num_base_bdevs_operational": 4, 00:11:21.842 "base_bdevs_list": [ 00:11:21.842 { 00:11:21.842 "name": "BaseBdev1", 00:11:21.842 "uuid": "c4dfb73c-2cae-4b09-ade8-2e61bf9143e0", 00:11:21.842 "is_configured": true, 00:11:21.842 "data_offset": 0, 00:11:21.842 "data_size": 65536 00:11:21.842 }, 00:11:21.842 { 00:11:21.842 "name": "BaseBdev2", 00:11:21.842 "uuid": "f6ed7921-d6cf-4973-b8f9-59c67a038d1f", 00:11:21.842 "is_configured": true, 00:11:21.842 "data_offset": 0, 00:11:21.842 "data_size": 65536 00:11:21.842 }, 00:11:21.842 { 00:11:21.842 "name": "BaseBdev3", 00:11:21.842 "uuid": "4f06b301-44df-464c-801f-aa1c09193a96", 00:11:21.842 "is_configured": true, 00:11:21.842 "data_offset": 0, 00:11:21.842 "data_size": 65536 00:11:21.842 }, 00:11:21.842 { 00:11:21.842 "name": "BaseBdev4", 00:11:21.842 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.842 "is_configured": false, 00:11:21.842 "data_offset": 0, 00:11:21.842 "data_size": 0 00:11:21.842 } 00:11:21.842 ] 00:11:21.842 }' 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:21.842 15:59:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.104 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:22.104 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.104 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.104 [2024-11-20 15:59:20.235042] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:22.104 [2024-11-20 15:59:20.235086] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:11:22.104 [2024-11-20 15:59:20.235095] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:11:22.104 [2024-11-20 15:59:20.235354] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:11:22.104 [2024-11-20 15:59:20.235503] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:11:22.104 [2024-11-20 15:59:20.235521] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:11:22.104 [2024-11-20 15:59:20.235748] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:22.104 BaseBdev4 00:11:22.104 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.104 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:11:22.104 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:11:22.104 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:22.104 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:22.104 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:22.104 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:22.104 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:22.104 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.104 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.104 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.104 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:22.104 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.104 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.104 [ 00:11:22.104 { 00:11:22.104 "name": "BaseBdev4", 00:11:22.104 "aliases": [ 00:11:22.104 "6b69aa5e-cda5-4163-aa69-ba7b6e554f62" 00:11:22.104 ], 00:11:22.104 "product_name": "Malloc disk", 00:11:22.104 "block_size": 512, 00:11:22.104 "num_blocks": 65536, 00:11:22.104 "uuid": "6b69aa5e-cda5-4163-aa69-ba7b6e554f62", 00:11:22.104 "assigned_rate_limits": { 00:11:22.104 "rw_ios_per_sec": 0, 00:11:22.104 "rw_mbytes_per_sec": 0, 00:11:22.104 "r_mbytes_per_sec": 0, 00:11:22.104 "w_mbytes_per_sec": 0 00:11:22.104 }, 00:11:22.104 "claimed": true, 00:11:22.104 "claim_type": "exclusive_write", 00:11:22.104 "zoned": false, 00:11:22.104 "supported_io_types": { 00:11:22.104 "read": true, 00:11:22.104 "write": true, 00:11:22.104 "unmap": true, 00:11:22.104 "flush": true, 00:11:22.104 "reset": true, 00:11:22.104 "nvme_admin": false, 00:11:22.104 "nvme_io": false, 00:11:22.104 "nvme_io_md": false, 00:11:22.104 "write_zeroes": true, 00:11:22.104 "zcopy": true, 00:11:22.104 "get_zone_info": false, 00:11:22.104 "zone_management": false, 00:11:22.104 "zone_append": false, 00:11:22.104 "compare": false, 00:11:22.104 "compare_and_write": false, 00:11:22.104 "abort": true, 00:11:22.104 "seek_hole": false, 00:11:22.104 "seek_data": false, 00:11:22.104 "copy": true, 00:11:22.105 "nvme_iov_md": false 00:11:22.105 }, 00:11:22.105 "memory_domains": [ 00:11:22.105 { 00:11:22.105 "dma_device_id": "system", 00:11:22.105 "dma_device_type": 1 00:11:22.105 }, 00:11:22.105 { 00:11:22.105 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:22.105 "dma_device_type": 2 00:11:22.105 } 00:11:22.105 ], 00:11:22.105 "driver_specific": {} 00:11:22.105 } 00:11:22.105 ] 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:22.105 "name": "Existed_Raid", 00:11:22.105 "uuid": "52c1115b-7839-4066-be34-d5341da94634", 00:11:22.105 "strip_size_kb": 64, 00:11:22.105 "state": "online", 00:11:22.105 "raid_level": "concat", 00:11:22.105 "superblock": false, 00:11:22.105 "num_base_bdevs": 4, 00:11:22.105 "num_base_bdevs_discovered": 4, 00:11:22.105 "num_base_bdevs_operational": 4, 00:11:22.105 "base_bdevs_list": [ 00:11:22.105 { 00:11:22.105 "name": "BaseBdev1", 00:11:22.105 "uuid": "c4dfb73c-2cae-4b09-ade8-2e61bf9143e0", 00:11:22.105 "is_configured": true, 00:11:22.105 "data_offset": 0, 00:11:22.105 "data_size": 65536 00:11:22.105 }, 00:11:22.105 { 00:11:22.105 "name": "BaseBdev2", 00:11:22.105 "uuid": "f6ed7921-d6cf-4973-b8f9-59c67a038d1f", 00:11:22.105 "is_configured": true, 00:11:22.105 "data_offset": 0, 00:11:22.105 "data_size": 65536 00:11:22.105 }, 00:11:22.105 { 00:11:22.105 "name": "BaseBdev3", 00:11:22.105 "uuid": "4f06b301-44df-464c-801f-aa1c09193a96", 00:11:22.105 "is_configured": true, 00:11:22.105 "data_offset": 0, 00:11:22.105 "data_size": 65536 00:11:22.105 }, 00:11:22.105 { 00:11:22.105 "name": "BaseBdev4", 00:11:22.105 "uuid": "6b69aa5e-cda5-4163-aa69-ba7b6e554f62", 00:11:22.105 "is_configured": true, 00:11:22.105 "data_offset": 0, 00:11:22.105 "data_size": 65536 00:11:22.105 } 00:11:22.105 ] 00:11:22.105 }' 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:22.105 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.366 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:11:22.366 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:22.366 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:22.366 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:22.366 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:22.366 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:22.366 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:22.366 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:22.366 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.366 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.366 [2024-11-20 15:59:20.599530] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:22.628 "name": "Existed_Raid", 00:11:22.628 "aliases": [ 00:11:22.628 "52c1115b-7839-4066-be34-d5341da94634" 00:11:22.628 ], 00:11:22.628 "product_name": "Raid Volume", 00:11:22.628 "block_size": 512, 00:11:22.628 "num_blocks": 262144, 00:11:22.628 "uuid": "52c1115b-7839-4066-be34-d5341da94634", 00:11:22.628 "assigned_rate_limits": { 00:11:22.628 "rw_ios_per_sec": 0, 00:11:22.628 "rw_mbytes_per_sec": 0, 00:11:22.628 "r_mbytes_per_sec": 0, 00:11:22.628 "w_mbytes_per_sec": 0 00:11:22.628 }, 00:11:22.628 "claimed": false, 00:11:22.628 "zoned": false, 00:11:22.628 "supported_io_types": { 00:11:22.628 "read": true, 00:11:22.628 "write": true, 00:11:22.628 "unmap": true, 00:11:22.628 "flush": true, 00:11:22.628 "reset": true, 00:11:22.628 "nvme_admin": false, 00:11:22.628 "nvme_io": false, 00:11:22.628 "nvme_io_md": false, 00:11:22.628 "write_zeroes": true, 00:11:22.628 "zcopy": false, 00:11:22.628 "get_zone_info": false, 00:11:22.628 "zone_management": false, 00:11:22.628 "zone_append": false, 00:11:22.628 "compare": false, 00:11:22.628 "compare_and_write": false, 00:11:22.628 "abort": false, 00:11:22.628 "seek_hole": false, 00:11:22.628 "seek_data": false, 00:11:22.628 "copy": false, 00:11:22.628 "nvme_iov_md": false 00:11:22.628 }, 00:11:22.628 "memory_domains": [ 00:11:22.628 { 00:11:22.628 "dma_device_id": "system", 00:11:22.628 "dma_device_type": 1 00:11:22.628 }, 00:11:22.628 { 00:11:22.628 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:22.628 "dma_device_type": 2 00:11:22.628 }, 00:11:22.628 { 00:11:22.628 "dma_device_id": "system", 00:11:22.628 "dma_device_type": 1 00:11:22.628 }, 00:11:22.628 { 00:11:22.628 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:22.628 "dma_device_type": 2 00:11:22.628 }, 00:11:22.628 { 00:11:22.628 "dma_device_id": "system", 00:11:22.628 "dma_device_type": 1 00:11:22.628 }, 00:11:22.628 { 00:11:22.628 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:22.628 "dma_device_type": 2 00:11:22.628 }, 00:11:22.628 { 00:11:22.628 "dma_device_id": "system", 00:11:22.628 "dma_device_type": 1 00:11:22.628 }, 00:11:22.628 { 00:11:22.628 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:22.628 "dma_device_type": 2 00:11:22.628 } 00:11:22.628 ], 00:11:22.628 "driver_specific": { 00:11:22.628 "raid": { 00:11:22.628 "uuid": "52c1115b-7839-4066-be34-d5341da94634", 00:11:22.628 "strip_size_kb": 64, 00:11:22.628 "state": "online", 00:11:22.628 "raid_level": "concat", 00:11:22.628 "superblock": false, 00:11:22.628 "num_base_bdevs": 4, 00:11:22.628 "num_base_bdevs_discovered": 4, 00:11:22.628 "num_base_bdevs_operational": 4, 00:11:22.628 "base_bdevs_list": [ 00:11:22.628 { 00:11:22.628 "name": "BaseBdev1", 00:11:22.628 "uuid": "c4dfb73c-2cae-4b09-ade8-2e61bf9143e0", 00:11:22.628 "is_configured": true, 00:11:22.628 "data_offset": 0, 00:11:22.628 "data_size": 65536 00:11:22.628 }, 00:11:22.628 { 00:11:22.628 "name": "BaseBdev2", 00:11:22.628 "uuid": "f6ed7921-d6cf-4973-b8f9-59c67a038d1f", 00:11:22.628 "is_configured": true, 00:11:22.628 "data_offset": 0, 00:11:22.628 "data_size": 65536 00:11:22.628 }, 00:11:22.628 { 00:11:22.628 "name": "BaseBdev3", 00:11:22.628 "uuid": "4f06b301-44df-464c-801f-aa1c09193a96", 00:11:22.628 "is_configured": true, 00:11:22.628 "data_offset": 0, 00:11:22.628 "data_size": 65536 00:11:22.628 }, 00:11:22.628 { 00:11:22.628 "name": "BaseBdev4", 00:11:22.628 "uuid": "6b69aa5e-cda5-4163-aa69-ba7b6e554f62", 00:11:22.628 "is_configured": true, 00:11:22.628 "data_offset": 0, 00:11:22.628 "data_size": 65536 00:11:22.628 } 00:11:22.628 ] 00:11:22.628 } 00:11:22.628 } 00:11:22.628 }' 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:11:22.628 BaseBdev2 00:11:22.628 BaseBdev3 00:11:22.628 BaseBdev4' 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.628 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.628 [2024-11-20 15:59:20.847260] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:22.628 [2024-11-20 15:59:20.847288] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:22.628 [2024-11-20 15:59:20.847333] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.890 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:22.890 "name": "Existed_Raid", 00:11:22.890 "uuid": "52c1115b-7839-4066-be34-d5341da94634", 00:11:22.890 "strip_size_kb": 64, 00:11:22.890 "state": "offline", 00:11:22.890 "raid_level": "concat", 00:11:22.890 "superblock": false, 00:11:22.890 "num_base_bdevs": 4, 00:11:22.890 "num_base_bdevs_discovered": 3, 00:11:22.890 "num_base_bdevs_operational": 3, 00:11:22.890 "base_bdevs_list": [ 00:11:22.890 { 00:11:22.890 "name": null, 00:11:22.890 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:22.890 "is_configured": false, 00:11:22.890 "data_offset": 0, 00:11:22.890 "data_size": 65536 00:11:22.890 }, 00:11:22.890 { 00:11:22.890 "name": "BaseBdev2", 00:11:22.891 "uuid": "f6ed7921-d6cf-4973-b8f9-59c67a038d1f", 00:11:22.891 "is_configured": true, 00:11:22.891 "data_offset": 0, 00:11:22.891 "data_size": 65536 00:11:22.891 }, 00:11:22.891 { 00:11:22.891 "name": "BaseBdev3", 00:11:22.891 "uuid": "4f06b301-44df-464c-801f-aa1c09193a96", 00:11:22.891 "is_configured": true, 00:11:22.891 "data_offset": 0, 00:11:22.891 "data_size": 65536 00:11:22.891 }, 00:11:22.891 { 00:11:22.891 "name": "BaseBdev4", 00:11:22.891 "uuid": "6b69aa5e-cda5-4163-aa69-ba7b6e554f62", 00:11:22.891 "is_configured": true, 00:11:22.891 "data_offset": 0, 00:11:22.891 "data_size": 65536 00:11:22.891 } 00:11:22.891 ] 00:11:22.891 }' 00:11:22.891 15:59:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:22.891 15:59:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.152 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:11:23.152 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.153 [2024-11-20 15:59:21.285731] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.153 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.153 [2024-11-20 15:59:21.384991] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.413 [2024-11-20 15:59:21.483143] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:11:23.413 [2024-11-20 15:59:21.483189] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.413 BaseBdev2 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.413 [ 00:11:23.413 { 00:11:23.413 "name": "BaseBdev2", 00:11:23.413 "aliases": [ 00:11:23.413 "b316acd1-60d9-4b15-a40e-8060dee6b216" 00:11:23.413 ], 00:11:23.413 "product_name": "Malloc disk", 00:11:23.413 "block_size": 512, 00:11:23.413 "num_blocks": 65536, 00:11:23.413 "uuid": "b316acd1-60d9-4b15-a40e-8060dee6b216", 00:11:23.413 "assigned_rate_limits": { 00:11:23.413 "rw_ios_per_sec": 0, 00:11:23.413 "rw_mbytes_per_sec": 0, 00:11:23.413 "r_mbytes_per_sec": 0, 00:11:23.413 "w_mbytes_per_sec": 0 00:11:23.413 }, 00:11:23.413 "claimed": false, 00:11:23.413 "zoned": false, 00:11:23.413 "supported_io_types": { 00:11:23.413 "read": true, 00:11:23.413 "write": true, 00:11:23.413 "unmap": true, 00:11:23.413 "flush": true, 00:11:23.413 "reset": true, 00:11:23.413 "nvme_admin": false, 00:11:23.413 "nvme_io": false, 00:11:23.413 "nvme_io_md": false, 00:11:23.413 "write_zeroes": true, 00:11:23.413 "zcopy": true, 00:11:23.413 "get_zone_info": false, 00:11:23.413 "zone_management": false, 00:11:23.413 "zone_append": false, 00:11:23.413 "compare": false, 00:11:23.413 "compare_and_write": false, 00:11:23.413 "abort": true, 00:11:23.413 "seek_hole": false, 00:11:23.413 "seek_data": false, 00:11:23.413 "copy": true, 00:11:23.413 "nvme_iov_md": false 00:11:23.413 }, 00:11:23.413 "memory_domains": [ 00:11:23.413 { 00:11:23.413 "dma_device_id": "system", 00:11:23.413 "dma_device_type": 1 00:11:23.413 }, 00:11:23.413 { 00:11:23.413 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:23.413 "dma_device_type": 2 00:11:23.413 } 00:11:23.413 ], 00:11:23.413 "driver_specific": {} 00:11:23.413 } 00:11:23.413 ] 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.413 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.696 BaseBdev3 00:11:23.696 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.696 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:11:23.696 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:11:23.696 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:23.696 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:23.696 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:23.696 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:23.696 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:23.696 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.696 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.696 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.696 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:23.696 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.696 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.696 [ 00:11:23.696 { 00:11:23.696 "name": "BaseBdev3", 00:11:23.696 "aliases": [ 00:11:23.696 "2eb3af20-3f1d-4a94-9908-ca8e2875bfad" 00:11:23.696 ], 00:11:23.696 "product_name": "Malloc disk", 00:11:23.696 "block_size": 512, 00:11:23.696 "num_blocks": 65536, 00:11:23.696 "uuid": "2eb3af20-3f1d-4a94-9908-ca8e2875bfad", 00:11:23.696 "assigned_rate_limits": { 00:11:23.696 "rw_ios_per_sec": 0, 00:11:23.696 "rw_mbytes_per_sec": 0, 00:11:23.696 "r_mbytes_per_sec": 0, 00:11:23.696 "w_mbytes_per_sec": 0 00:11:23.696 }, 00:11:23.696 "claimed": false, 00:11:23.696 "zoned": false, 00:11:23.696 "supported_io_types": { 00:11:23.696 "read": true, 00:11:23.696 "write": true, 00:11:23.696 "unmap": true, 00:11:23.696 "flush": true, 00:11:23.696 "reset": true, 00:11:23.696 "nvme_admin": false, 00:11:23.696 "nvme_io": false, 00:11:23.696 "nvme_io_md": false, 00:11:23.696 "write_zeroes": true, 00:11:23.696 "zcopy": true, 00:11:23.696 "get_zone_info": false, 00:11:23.697 "zone_management": false, 00:11:23.697 "zone_append": false, 00:11:23.697 "compare": false, 00:11:23.697 "compare_and_write": false, 00:11:23.697 "abort": true, 00:11:23.697 "seek_hole": false, 00:11:23.697 "seek_data": false, 00:11:23.697 "copy": true, 00:11:23.697 "nvme_iov_md": false 00:11:23.697 }, 00:11:23.697 "memory_domains": [ 00:11:23.697 { 00:11:23.697 "dma_device_id": "system", 00:11:23.697 "dma_device_type": 1 00:11:23.697 }, 00:11:23.697 { 00:11:23.697 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:23.697 "dma_device_type": 2 00:11:23.697 } 00:11:23.697 ], 00:11:23.697 "driver_specific": {} 00:11:23.697 } 00:11:23.697 ] 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.697 BaseBdev4 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.697 [ 00:11:23.697 { 00:11:23.697 "name": "BaseBdev4", 00:11:23.697 "aliases": [ 00:11:23.697 "d45a0d40-41a8-4501-8838-83f8ebf6d17e" 00:11:23.697 ], 00:11:23.697 "product_name": "Malloc disk", 00:11:23.697 "block_size": 512, 00:11:23.697 "num_blocks": 65536, 00:11:23.697 "uuid": "d45a0d40-41a8-4501-8838-83f8ebf6d17e", 00:11:23.697 "assigned_rate_limits": { 00:11:23.697 "rw_ios_per_sec": 0, 00:11:23.697 "rw_mbytes_per_sec": 0, 00:11:23.697 "r_mbytes_per_sec": 0, 00:11:23.697 "w_mbytes_per_sec": 0 00:11:23.697 }, 00:11:23.697 "claimed": false, 00:11:23.697 "zoned": false, 00:11:23.697 "supported_io_types": { 00:11:23.697 "read": true, 00:11:23.697 "write": true, 00:11:23.697 "unmap": true, 00:11:23.697 "flush": true, 00:11:23.697 "reset": true, 00:11:23.697 "nvme_admin": false, 00:11:23.697 "nvme_io": false, 00:11:23.697 "nvme_io_md": false, 00:11:23.697 "write_zeroes": true, 00:11:23.697 "zcopy": true, 00:11:23.697 "get_zone_info": false, 00:11:23.697 "zone_management": false, 00:11:23.697 "zone_append": false, 00:11:23.697 "compare": false, 00:11:23.697 "compare_and_write": false, 00:11:23.697 "abort": true, 00:11:23.697 "seek_hole": false, 00:11:23.697 "seek_data": false, 00:11:23.697 "copy": true, 00:11:23.697 "nvme_iov_md": false 00:11:23.697 }, 00:11:23.697 "memory_domains": [ 00:11:23.697 { 00:11:23.697 "dma_device_id": "system", 00:11:23.697 "dma_device_type": 1 00:11:23.697 }, 00:11:23.697 { 00:11:23.697 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:23.697 "dma_device_type": 2 00:11:23.697 } 00:11:23.697 ], 00:11:23.697 "driver_specific": {} 00:11:23.697 } 00:11:23.697 ] 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.697 [2024-11-20 15:59:21.765604] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:23.697 [2024-11-20 15:59:21.765643] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:23.697 [2024-11-20 15:59:21.765664] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:23.697 [2024-11-20 15:59:21.767490] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:23.697 [2024-11-20 15:59:21.767543] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:23.697 "name": "Existed_Raid", 00:11:23.697 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:23.697 "strip_size_kb": 64, 00:11:23.697 "state": "configuring", 00:11:23.697 "raid_level": "concat", 00:11:23.697 "superblock": false, 00:11:23.697 "num_base_bdevs": 4, 00:11:23.697 "num_base_bdevs_discovered": 3, 00:11:23.697 "num_base_bdevs_operational": 4, 00:11:23.697 "base_bdevs_list": [ 00:11:23.697 { 00:11:23.697 "name": "BaseBdev1", 00:11:23.697 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:23.697 "is_configured": false, 00:11:23.697 "data_offset": 0, 00:11:23.697 "data_size": 0 00:11:23.697 }, 00:11:23.697 { 00:11:23.697 "name": "BaseBdev2", 00:11:23.697 "uuid": "b316acd1-60d9-4b15-a40e-8060dee6b216", 00:11:23.697 "is_configured": true, 00:11:23.697 "data_offset": 0, 00:11:23.697 "data_size": 65536 00:11:23.697 }, 00:11:23.697 { 00:11:23.697 "name": "BaseBdev3", 00:11:23.697 "uuid": "2eb3af20-3f1d-4a94-9908-ca8e2875bfad", 00:11:23.697 "is_configured": true, 00:11:23.697 "data_offset": 0, 00:11:23.697 "data_size": 65536 00:11:23.697 }, 00:11:23.697 { 00:11:23.697 "name": "BaseBdev4", 00:11:23.697 "uuid": "d45a0d40-41a8-4501-8838-83f8ebf6d17e", 00:11:23.697 "is_configured": true, 00:11:23.697 "data_offset": 0, 00:11:23.697 "data_size": 65536 00:11:23.697 } 00:11:23.697 ] 00:11:23.697 }' 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:23.697 15:59:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.976 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:11:23.976 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.976 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.976 [2024-11-20 15:59:22.089699] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:23.976 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.976 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:23.976 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:23.976 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:23.976 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:23.976 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:23.976 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:23.976 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:23.976 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:23.976 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:23.976 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:23.976 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:23.976 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.976 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.977 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.977 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.977 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:23.977 "name": "Existed_Raid", 00:11:23.977 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:23.977 "strip_size_kb": 64, 00:11:23.977 "state": "configuring", 00:11:23.977 "raid_level": "concat", 00:11:23.977 "superblock": false, 00:11:23.977 "num_base_bdevs": 4, 00:11:23.977 "num_base_bdevs_discovered": 2, 00:11:23.977 "num_base_bdevs_operational": 4, 00:11:23.977 "base_bdevs_list": [ 00:11:23.977 { 00:11:23.977 "name": "BaseBdev1", 00:11:23.977 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:23.977 "is_configured": false, 00:11:23.977 "data_offset": 0, 00:11:23.977 "data_size": 0 00:11:23.977 }, 00:11:23.977 { 00:11:23.977 "name": null, 00:11:23.977 "uuid": "b316acd1-60d9-4b15-a40e-8060dee6b216", 00:11:23.977 "is_configured": false, 00:11:23.977 "data_offset": 0, 00:11:23.977 "data_size": 65536 00:11:23.977 }, 00:11:23.977 { 00:11:23.977 "name": "BaseBdev3", 00:11:23.977 "uuid": "2eb3af20-3f1d-4a94-9908-ca8e2875bfad", 00:11:23.977 "is_configured": true, 00:11:23.977 "data_offset": 0, 00:11:23.977 "data_size": 65536 00:11:23.977 }, 00:11:23.977 { 00:11:23.977 "name": "BaseBdev4", 00:11:23.977 "uuid": "d45a0d40-41a8-4501-8838-83f8ebf6d17e", 00:11:23.977 "is_configured": true, 00:11:23.977 "data_offset": 0, 00:11:23.977 "data_size": 65536 00:11:23.977 } 00:11:23.977 ] 00:11:23.977 }' 00:11:23.977 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:23.977 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.237 [2024-11-20 15:59:22.476417] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:24.237 BaseBdev1 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.237 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.498 [ 00:11:24.498 { 00:11:24.498 "name": "BaseBdev1", 00:11:24.498 "aliases": [ 00:11:24.498 "57b023b4-39a9-41d1-9f87-6166c78c372c" 00:11:24.498 ], 00:11:24.498 "product_name": "Malloc disk", 00:11:24.498 "block_size": 512, 00:11:24.498 "num_blocks": 65536, 00:11:24.498 "uuid": "57b023b4-39a9-41d1-9f87-6166c78c372c", 00:11:24.498 "assigned_rate_limits": { 00:11:24.498 "rw_ios_per_sec": 0, 00:11:24.498 "rw_mbytes_per_sec": 0, 00:11:24.498 "r_mbytes_per_sec": 0, 00:11:24.498 "w_mbytes_per_sec": 0 00:11:24.498 }, 00:11:24.498 "claimed": true, 00:11:24.498 "claim_type": "exclusive_write", 00:11:24.498 "zoned": false, 00:11:24.498 "supported_io_types": { 00:11:24.498 "read": true, 00:11:24.498 "write": true, 00:11:24.498 "unmap": true, 00:11:24.498 "flush": true, 00:11:24.498 "reset": true, 00:11:24.498 "nvme_admin": false, 00:11:24.498 "nvme_io": false, 00:11:24.498 "nvme_io_md": false, 00:11:24.498 "write_zeroes": true, 00:11:24.498 "zcopy": true, 00:11:24.498 "get_zone_info": false, 00:11:24.498 "zone_management": false, 00:11:24.498 "zone_append": false, 00:11:24.498 "compare": false, 00:11:24.498 "compare_and_write": false, 00:11:24.498 "abort": true, 00:11:24.498 "seek_hole": false, 00:11:24.498 "seek_data": false, 00:11:24.498 "copy": true, 00:11:24.498 "nvme_iov_md": false 00:11:24.498 }, 00:11:24.498 "memory_domains": [ 00:11:24.498 { 00:11:24.498 "dma_device_id": "system", 00:11:24.498 "dma_device_type": 1 00:11:24.498 }, 00:11:24.498 { 00:11:24.498 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:24.498 "dma_device_type": 2 00:11:24.498 } 00:11:24.498 ], 00:11:24.498 "driver_specific": {} 00:11:24.498 } 00:11:24.498 ] 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:24.498 "name": "Existed_Raid", 00:11:24.498 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:24.498 "strip_size_kb": 64, 00:11:24.498 "state": "configuring", 00:11:24.498 "raid_level": "concat", 00:11:24.498 "superblock": false, 00:11:24.498 "num_base_bdevs": 4, 00:11:24.498 "num_base_bdevs_discovered": 3, 00:11:24.498 "num_base_bdevs_operational": 4, 00:11:24.498 "base_bdevs_list": [ 00:11:24.498 { 00:11:24.498 "name": "BaseBdev1", 00:11:24.498 "uuid": "57b023b4-39a9-41d1-9f87-6166c78c372c", 00:11:24.498 "is_configured": true, 00:11:24.498 "data_offset": 0, 00:11:24.498 "data_size": 65536 00:11:24.498 }, 00:11:24.498 { 00:11:24.498 "name": null, 00:11:24.498 "uuid": "b316acd1-60d9-4b15-a40e-8060dee6b216", 00:11:24.498 "is_configured": false, 00:11:24.498 "data_offset": 0, 00:11:24.498 "data_size": 65536 00:11:24.498 }, 00:11:24.498 { 00:11:24.498 "name": "BaseBdev3", 00:11:24.498 "uuid": "2eb3af20-3f1d-4a94-9908-ca8e2875bfad", 00:11:24.498 "is_configured": true, 00:11:24.498 "data_offset": 0, 00:11:24.498 "data_size": 65536 00:11:24.498 }, 00:11:24.498 { 00:11:24.498 "name": "BaseBdev4", 00:11:24.498 "uuid": "d45a0d40-41a8-4501-8838-83f8ebf6d17e", 00:11:24.498 "is_configured": true, 00:11:24.498 "data_offset": 0, 00:11:24.498 "data_size": 65536 00:11:24.498 } 00:11:24.498 ] 00:11:24.498 }' 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:24.498 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.759 [2024-11-20 15:59:22.868600] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:24.759 "name": "Existed_Raid", 00:11:24.759 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:24.759 "strip_size_kb": 64, 00:11:24.759 "state": "configuring", 00:11:24.759 "raid_level": "concat", 00:11:24.759 "superblock": false, 00:11:24.759 "num_base_bdevs": 4, 00:11:24.759 "num_base_bdevs_discovered": 2, 00:11:24.759 "num_base_bdevs_operational": 4, 00:11:24.759 "base_bdevs_list": [ 00:11:24.759 { 00:11:24.759 "name": "BaseBdev1", 00:11:24.759 "uuid": "57b023b4-39a9-41d1-9f87-6166c78c372c", 00:11:24.759 "is_configured": true, 00:11:24.759 "data_offset": 0, 00:11:24.759 "data_size": 65536 00:11:24.759 }, 00:11:24.759 { 00:11:24.759 "name": null, 00:11:24.759 "uuid": "b316acd1-60d9-4b15-a40e-8060dee6b216", 00:11:24.759 "is_configured": false, 00:11:24.759 "data_offset": 0, 00:11:24.759 "data_size": 65536 00:11:24.759 }, 00:11:24.759 { 00:11:24.759 "name": null, 00:11:24.759 "uuid": "2eb3af20-3f1d-4a94-9908-ca8e2875bfad", 00:11:24.759 "is_configured": false, 00:11:24.759 "data_offset": 0, 00:11:24.759 "data_size": 65536 00:11:24.759 }, 00:11:24.759 { 00:11:24.759 "name": "BaseBdev4", 00:11:24.759 "uuid": "d45a0d40-41a8-4501-8838-83f8ebf6d17e", 00:11:24.759 "is_configured": true, 00:11:24.759 "data_offset": 0, 00:11:24.759 "data_size": 65536 00:11:24.759 } 00:11:24.759 ] 00:11:24.759 }' 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:24.759 15:59:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.021 [2024-11-20 15:59:23.228689] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:25.021 "name": "Existed_Raid", 00:11:25.021 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:25.021 "strip_size_kb": 64, 00:11:25.021 "state": "configuring", 00:11:25.021 "raid_level": "concat", 00:11:25.021 "superblock": false, 00:11:25.021 "num_base_bdevs": 4, 00:11:25.021 "num_base_bdevs_discovered": 3, 00:11:25.021 "num_base_bdevs_operational": 4, 00:11:25.021 "base_bdevs_list": [ 00:11:25.021 { 00:11:25.021 "name": "BaseBdev1", 00:11:25.021 "uuid": "57b023b4-39a9-41d1-9f87-6166c78c372c", 00:11:25.021 "is_configured": true, 00:11:25.021 "data_offset": 0, 00:11:25.021 "data_size": 65536 00:11:25.021 }, 00:11:25.021 { 00:11:25.021 "name": null, 00:11:25.021 "uuid": "b316acd1-60d9-4b15-a40e-8060dee6b216", 00:11:25.021 "is_configured": false, 00:11:25.021 "data_offset": 0, 00:11:25.021 "data_size": 65536 00:11:25.021 }, 00:11:25.021 { 00:11:25.021 "name": "BaseBdev3", 00:11:25.021 "uuid": "2eb3af20-3f1d-4a94-9908-ca8e2875bfad", 00:11:25.021 "is_configured": true, 00:11:25.021 "data_offset": 0, 00:11:25.021 "data_size": 65536 00:11:25.021 }, 00:11:25.021 { 00:11:25.021 "name": "BaseBdev4", 00:11:25.021 "uuid": "d45a0d40-41a8-4501-8838-83f8ebf6d17e", 00:11:25.021 "is_configured": true, 00:11:25.021 "data_offset": 0, 00:11:25.021 "data_size": 65536 00:11:25.021 } 00:11:25.021 ] 00:11:25.021 }' 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:25.021 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.590 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.591 [2024-11-20 15:59:23.584793] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:25.591 "name": "Existed_Raid", 00:11:25.591 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:25.591 "strip_size_kb": 64, 00:11:25.591 "state": "configuring", 00:11:25.591 "raid_level": "concat", 00:11:25.591 "superblock": false, 00:11:25.591 "num_base_bdevs": 4, 00:11:25.591 "num_base_bdevs_discovered": 2, 00:11:25.591 "num_base_bdevs_operational": 4, 00:11:25.591 "base_bdevs_list": [ 00:11:25.591 { 00:11:25.591 "name": null, 00:11:25.591 "uuid": "57b023b4-39a9-41d1-9f87-6166c78c372c", 00:11:25.591 "is_configured": false, 00:11:25.591 "data_offset": 0, 00:11:25.591 "data_size": 65536 00:11:25.591 }, 00:11:25.591 { 00:11:25.591 "name": null, 00:11:25.591 "uuid": "b316acd1-60d9-4b15-a40e-8060dee6b216", 00:11:25.591 "is_configured": false, 00:11:25.591 "data_offset": 0, 00:11:25.591 "data_size": 65536 00:11:25.591 }, 00:11:25.591 { 00:11:25.591 "name": "BaseBdev3", 00:11:25.591 "uuid": "2eb3af20-3f1d-4a94-9908-ca8e2875bfad", 00:11:25.591 "is_configured": true, 00:11:25.591 "data_offset": 0, 00:11:25.591 "data_size": 65536 00:11:25.591 }, 00:11:25.591 { 00:11:25.591 "name": "BaseBdev4", 00:11:25.591 "uuid": "d45a0d40-41a8-4501-8838-83f8ebf6d17e", 00:11:25.591 "is_configured": true, 00:11:25.591 "data_offset": 0, 00:11:25.591 "data_size": 65536 00:11:25.591 } 00:11:25.591 ] 00:11:25.591 }' 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:25.591 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.848 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.848 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.848 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.849 [2024-11-20 15:59:23.983676] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.849 15:59:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.849 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.849 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:25.849 "name": "Existed_Raid", 00:11:25.849 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:25.849 "strip_size_kb": 64, 00:11:25.849 "state": "configuring", 00:11:25.849 "raid_level": "concat", 00:11:25.849 "superblock": false, 00:11:25.849 "num_base_bdevs": 4, 00:11:25.849 "num_base_bdevs_discovered": 3, 00:11:25.849 "num_base_bdevs_operational": 4, 00:11:25.849 "base_bdevs_list": [ 00:11:25.849 { 00:11:25.849 "name": null, 00:11:25.849 "uuid": "57b023b4-39a9-41d1-9f87-6166c78c372c", 00:11:25.849 "is_configured": false, 00:11:25.849 "data_offset": 0, 00:11:25.849 "data_size": 65536 00:11:25.849 }, 00:11:25.849 { 00:11:25.849 "name": "BaseBdev2", 00:11:25.849 "uuid": "b316acd1-60d9-4b15-a40e-8060dee6b216", 00:11:25.849 "is_configured": true, 00:11:25.849 "data_offset": 0, 00:11:25.849 "data_size": 65536 00:11:25.849 }, 00:11:25.849 { 00:11:25.849 "name": "BaseBdev3", 00:11:25.849 "uuid": "2eb3af20-3f1d-4a94-9908-ca8e2875bfad", 00:11:25.849 "is_configured": true, 00:11:25.849 "data_offset": 0, 00:11:25.849 "data_size": 65536 00:11:25.849 }, 00:11:25.849 { 00:11:25.849 "name": "BaseBdev4", 00:11:25.849 "uuid": "d45a0d40-41a8-4501-8838-83f8ebf6d17e", 00:11:25.849 "is_configured": true, 00:11:25.849 "data_offset": 0, 00:11:25.849 "data_size": 65536 00:11:25.849 } 00:11:25.849 ] 00:11:25.849 }' 00:11:25.849 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:25.849 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.106 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:26.106 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:26.106 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.106 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.106 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.106 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:11:26.106 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:26.106 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.106 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.106 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:11:26.106 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.106 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 57b023b4-39a9-41d1-9f87-6166c78c372c 00:11:26.106 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.106 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.374 [2024-11-20 15:59:24.373737] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:11:26.374 [2024-11-20 15:59:24.373773] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:11:26.374 [2024-11-20 15:59:24.373779] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:11:26.374 [2024-11-20 15:59:24.373990] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:11:26.375 [2024-11-20 15:59:24.374091] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:11:26.375 [2024-11-20 15:59:24.374099] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:11:26.375 [2024-11-20 15:59:24.374266] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:26.375 NewBaseBdev 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.375 [ 00:11:26.375 { 00:11:26.375 "name": "NewBaseBdev", 00:11:26.375 "aliases": [ 00:11:26.375 "57b023b4-39a9-41d1-9f87-6166c78c372c" 00:11:26.375 ], 00:11:26.375 "product_name": "Malloc disk", 00:11:26.375 "block_size": 512, 00:11:26.375 "num_blocks": 65536, 00:11:26.375 "uuid": "57b023b4-39a9-41d1-9f87-6166c78c372c", 00:11:26.375 "assigned_rate_limits": { 00:11:26.375 "rw_ios_per_sec": 0, 00:11:26.375 "rw_mbytes_per_sec": 0, 00:11:26.375 "r_mbytes_per_sec": 0, 00:11:26.375 "w_mbytes_per_sec": 0 00:11:26.375 }, 00:11:26.375 "claimed": true, 00:11:26.375 "claim_type": "exclusive_write", 00:11:26.375 "zoned": false, 00:11:26.375 "supported_io_types": { 00:11:26.375 "read": true, 00:11:26.375 "write": true, 00:11:26.375 "unmap": true, 00:11:26.375 "flush": true, 00:11:26.375 "reset": true, 00:11:26.375 "nvme_admin": false, 00:11:26.375 "nvme_io": false, 00:11:26.375 "nvme_io_md": false, 00:11:26.375 "write_zeroes": true, 00:11:26.375 "zcopy": true, 00:11:26.375 "get_zone_info": false, 00:11:26.375 "zone_management": false, 00:11:26.375 "zone_append": false, 00:11:26.375 "compare": false, 00:11:26.375 "compare_and_write": false, 00:11:26.375 "abort": true, 00:11:26.375 "seek_hole": false, 00:11:26.375 "seek_data": false, 00:11:26.375 "copy": true, 00:11:26.375 "nvme_iov_md": false 00:11:26.375 }, 00:11:26.375 "memory_domains": [ 00:11:26.375 { 00:11:26.375 "dma_device_id": "system", 00:11:26.375 "dma_device_type": 1 00:11:26.375 }, 00:11:26.375 { 00:11:26.375 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:26.375 "dma_device_type": 2 00:11:26.375 } 00:11:26.375 ], 00:11:26.375 "driver_specific": {} 00:11:26.375 } 00:11:26.375 ] 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:26.375 "name": "Existed_Raid", 00:11:26.375 "uuid": "0cb97300-f4ab-44dd-8943-1011a6784f9f", 00:11:26.375 "strip_size_kb": 64, 00:11:26.375 "state": "online", 00:11:26.375 "raid_level": "concat", 00:11:26.375 "superblock": false, 00:11:26.375 "num_base_bdevs": 4, 00:11:26.375 "num_base_bdevs_discovered": 4, 00:11:26.375 "num_base_bdevs_operational": 4, 00:11:26.375 "base_bdevs_list": [ 00:11:26.375 { 00:11:26.375 "name": "NewBaseBdev", 00:11:26.375 "uuid": "57b023b4-39a9-41d1-9f87-6166c78c372c", 00:11:26.375 "is_configured": true, 00:11:26.375 "data_offset": 0, 00:11:26.375 "data_size": 65536 00:11:26.375 }, 00:11:26.375 { 00:11:26.375 "name": "BaseBdev2", 00:11:26.375 "uuid": "b316acd1-60d9-4b15-a40e-8060dee6b216", 00:11:26.375 "is_configured": true, 00:11:26.375 "data_offset": 0, 00:11:26.375 "data_size": 65536 00:11:26.375 }, 00:11:26.375 { 00:11:26.375 "name": "BaseBdev3", 00:11:26.375 "uuid": "2eb3af20-3f1d-4a94-9908-ca8e2875bfad", 00:11:26.375 "is_configured": true, 00:11:26.375 "data_offset": 0, 00:11:26.375 "data_size": 65536 00:11:26.375 }, 00:11:26.375 { 00:11:26.375 "name": "BaseBdev4", 00:11:26.375 "uuid": "d45a0d40-41a8-4501-8838-83f8ebf6d17e", 00:11:26.375 "is_configured": true, 00:11:26.375 "data_offset": 0, 00:11:26.375 "data_size": 65536 00:11:26.375 } 00:11:26.375 ] 00:11:26.375 }' 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:26.375 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.634 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:11:26.634 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:26.634 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:26.634 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:26.634 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:26.634 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:26.634 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:26.634 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:26.634 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.634 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.634 [2024-11-20 15:59:24.730152] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:26.634 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.634 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:26.634 "name": "Existed_Raid", 00:11:26.634 "aliases": [ 00:11:26.634 "0cb97300-f4ab-44dd-8943-1011a6784f9f" 00:11:26.634 ], 00:11:26.634 "product_name": "Raid Volume", 00:11:26.634 "block_size": 512, 00:11:26.634 "num_blocks": 262144, 00:11:26.634 "uuid": "0cb97300-f4ab-44dd-8943-1011a6784f9f", 00:11:26.634 "assigned_rate_limits": { 00:11:26.634 "rw_ios_per_sec": 0, 00:11:26.634 "rw_mbytes_per_sec": 0, 00:11:26.634 "r_mbytes_per_sec": 0, 00:11:26.634 "w_mbytes_per_sec": 0 00:11:26.634 }, 00:11:26.634 "claimed": false, 00:11:26.634 "zoned": false, 00:11:26.634 "supported_io_types": { 00:11:26.634 "read": true, 00:11:26.634 "write": true, 00:11:26.634 "unmap": true, 00:11:26.634 "flush": true, 00:11:26.634 "reset": true, 00:11:26.634 "nvme_admin": false, 00:11:26.634 "nvme_io": false, 00:11:26.634 "nvme_io_md": false, 00:11:26.634 "write_zeroes": true, 00:11:26.634 "zcopy": false, 00:11:26.634 "get_zone_info": false, 00:11:26.634 "zone_management": false, 00:11:26.634 "zone_append": false, 00:11:26.634 "compare": false, 00:11:26.634 "compare_and_write": false, 00:11:26.634 "abort": false, 00:11:26.634 "seek_hole": false, 00:11:26.634 "seek_data": false, 00:11:26.634 "copy": false, 00:11:26.634 "nvme_iov_md": false 00:11:26.634 }, 00:11:26.634 "memory_domains": [ 00:11:26.634 { 00:11:26.634 "dma_device_id": "system", 00:11:26.634 "dma_device_type": 1 00:11:26.634 }, 00:11:26.634 { 00:11:26.634 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:26.634 "dma_device_type": 2 00:11:26.634 }, 00:11:26.634 { 00:11:26.634 "dma_device_id": "system", 00:11:26.634 "dma_device_type": 1 00:11:26.634 }, 00:11:26.634 { 00:11:26.634 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:26.634 "dma_device_type": 2 00:11:26.634 }, 00:11:26.634 { 00:11:26.634 "dma_device_id": "system", 00:11:26.634 "dma_device_type": 1 00:11:26.634 }, 00:11:26.634 { 00:11:26.634 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:26.634 "dma_device_type": 2 00:11:26.634 }, 00:11:26.634 { 00:11:26.634 "dma_device_id": "system", 00:11:26.634 "dma_device_type": 1 00:11:26.634 }, 00:11:26.634 { 00:11:26.634 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:26.634 "dma_device_type": 2 00:11:26.634 } 00:11:26.634 ], 00:11:26.634 "driver_specific": { 00:11:26.634 "raid": { 00:11:26.634 "uuid": "0cb97300-f4ab-44dd-8943-1011a6784f9f", 00:11:26.634 "strip_size_kb": 64, 00:11:26.634 "state": "online", 00:11:26.634 "raid_level": "concat", 00:11:26.634 "superblock": false, 00:11:26.634 "num_base_bdevs": 4, 00:11:26.634 "num_base_bdevs_discovered": 4, 00:11:26.634 "num_base_bdevs_operational": 4, 00:11:26.634 "base_bdevs_list": [ 00:11:26.634 { 00:11:26.634 "name": "NewBaseBdev", 00:11:26.634 "uuid": "57b023b4-39a9-41d1-9f87-6166c78c372c", 00:11:26.634 "is_configured": true, 00:11:26.634 "data_offset": 0, 00:11:26.634 "data_size": 65536 00:11:26.634 }, 00:11:26.634 { 00:11:26.634 "name": "BaseBdev2", 00:11:26.634 "uuid": "b316acd1-60d9-4b15-a40e-8060dee6b216", 00:11:26.634 "is_configured": true, 00:11:26.634 "data_offset": 0, 00:11:26.634 "data_size": 65536 00:11:26.634 }, 00:11:26.634 { 00:11:26.634 "name": "BaseBdev3", 00:11:26.634 "uuid": "2eb3af20-3f1d-4a94-9908-ca8e2875bfad", 00:11:26.634 "is_configured": true, 00:11:26.634 "data_offset": 0, 00:11:26.634 "data_size": 65536 00:11:26.634 }, 00:11:26.634 { 00:11:26.634 "name": "BaseBdev4", 00:11:26.634 "uuid": "d45a0d40-41a8-4501-8838-83f8ebf6d17e", 00:11:26.634 "is_configured": true, 00:11:26.634 "data_offset": 0, 00:11:26.634 "data_size": 65536 00:11:26.634 } 00:11:26.634 ] 00:11:26.634 } 00:11:26.634 } 00:11:26.634 }' 00:11:26.634 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:26.634 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:11:26.634 BaseBdev2 00:11:26.634 BaseBdev3 00:11:26.634 BaseBdev4' 00:11:26.634 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:26.635 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:26.635 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:26.635 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:11:26.635 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.635 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.635 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:26.635 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.635 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:26.635 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:26.635 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:26.635 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:26.635 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:26.635 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.635 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.635 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.895 [2024-11-20 15:59:24.945864] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:26.895 [2024-11-20 15:59:24.945889] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:26.895 [2024-11-20 15:59:24.945943] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:26.895 [2024-11-20 15:59:24.945996] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:26.895 [2024-11-20 15:59:24.946004] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 69353 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 69353 ']' 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 69353 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69353 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:26.895 killing process with pid 69353 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69353' 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 69353 00:11:26.895 [2024-11-20 15:59:24.975594] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:26.895 15:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 69353 00:11:27.156 [2024-11-20 15:59:25.203685] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:27.801 ************************************ 00:11:27.801 END TEST raid_state_function_test 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:11:27.801 00:11:27.801 real 0m8.502s 00:11:27.801 user 0m13.478s 00:11:27.801 sys 0m1.405s 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.801 ************************************ 00:11:27.801 15:59:25 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 4 true 00:11:27.801 15:59:25 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:27.801 15:59:25 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:27.801 15:59:25 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:27.801 ************************************ 00:11:27.801 START TEST raid_state_function_test_sb 00:11:27.801 ************************************ 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 4 true 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:11:27.801 Process raid pid: 69997 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=69997 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 69997' 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 69997 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 69997 ']' 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:27.801 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.801 15:59:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:11:28.061 [2024-11-20 15:59:26.057368] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:11:28.061 [2024-11-20 15:59:26.057490] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:28.061 [2024-11-20 15:59:26.219976] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:28.322 [2024-11-20 15:59:26.326098] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:28.322 [2024-11-20 15:59:26.465788] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:28.322 [2024-11-20 15:59:26.465827] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.892 [2024-11-20 15:59:26.917317] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:28.892 [2024-11-20 15:59:26.917370] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:28.892 [2024-11-20 15:59:26.917386] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:28.892 [2024-11-20 15:59:26.917397] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:28.892 [2024-11-20 15:59:26.917404] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:28.892 [2024-11-20 15:59:26.917414] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:28.892 [2024-11-20 15:59:26.917420] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:28.892 [2024-11-20 15:59:26.917429] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:28.892 "name": "Existed_Raid", 00:11:28.892 "uuid": "fd9dcdc5-0925-4cfd-9cbf-7d853e90cf50", 00:11:28.892 "strip_size_kb": 64, 00:11:28.892 "state": "configuring", 00:11:28.892 "raid_level": "concat", 00:11:28.892 "superblock": true, 00:11:28.892 "num_base_bdevs": 4, 00:11:28.892 "num_base_bdevs_discovered": 0, 00:11:28.892 "num_base_bdevs_operational": 4, 00:11:28.892 "base_bdevs_list": [ 00:11:28.892 { 00:11:28.892 "name": "BaseBdev1", 00:11:28.892 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:28.892 "is_configured": false, 00:11:28.892 "data_offset": 0, 00:11:28.892 "data_size": 0 00:11:28.892 }, 00:11:28.892 { 00:11:28.892 "name": "BaseBdev2", 00:11:28.892 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:28.892 "is_configured": false, 00:11:28.892 "data_offset": 0, 00:11:28.892 "data_size": 0 00:11:28.892 }, 00:11:28.892 { 00:11:28.892 "name": "BaseBdev3", 00:11:28.892 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:28.892 "is_configured": false, 00:11:28.892 "data_offset": 0, 00:11:28.892 "data_size": 0 00:11:28.892 }, 00:11:28.892 { 00:11:28.892 "name": "BaseBdev4", 00:11:28.892 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:28.892 "is_configured": false, 00:11:28.892 "data_offset": 0, 00:11:28.892 "data_size": 0 00:11:28.892 } 00:11:28.892 ] 00:11:28.892 }' 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:28.892 15:59:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.152 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:29.152 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.152 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.152 [2024-11-20 15:59:27.249323] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:29.152 [2024-11-20 15:59:27.249356] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:11:29.152 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.152 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:29.152 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.152 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.152 [2024-11-20 15:59:27.257343] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:29.152 [2024-11-20 15:59:27.257377] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:29.152 [2024-11-20 15:59:27.257385] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:29.152 [2024-11-20 15:59:27.257393] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:29.152 [2024-11-20 15:59:27.257399] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:29.152 [2024-11-20 15:59:27.257408] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:29.152 [2024-11-20 15:59:27.257414] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:29.152 [2024-11-20 15:59:27.257422] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:29.152 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.152 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:29.152 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.152 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.152 [2024-11-20 15:59:27.289920] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:29.152 BaseBdev1 00:11:29.152 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.152 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:11:29.152 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:11:29.152 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:29.152 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.153 [ 00:11:29.153 { 00:11:29.153 "name": "BaseBdev1", 00:11:29.153 "aliases": [ 00:11:29.153 "cae646d7-92a7-42eb-a42e-9f81fb9e9c56" 00:11:29.153 ], 00:11:29.153 "product_name": "Malloc disk", 00:11:29.153 "block_size": 512, 00:11:29.153 "num_blocks": 65536, 00:11:29.153 "uuid": "cae646d7-92a7-42eb-a42e-9f81fb9e9c56", 00:11:29.153 "assigned_rate_limits": { 00:11:29.153 "rw_ios_per_sec": 0, 00:11:29.153 "rw_mbytes_per_sec": 0, 00:11:29.153 "r_mbytes_per_sec": 0, 00:11:29.153 "w_mbytes_per_sec": 0 00:11:29.153 }, 00:11:29.153 "claimed": true, 00:11:29.153 "claim_type": "exclusive_write", 00:11:29.153 "zoned": false, 00:11:29.153 "supported_io_types": { 00:11:29.153 "read": true, 00:11:29.153 "write": true, 00:11:29.153 "unmap": true, 00:11:29.153 "flush": true, 00:11:29.153 "reset": true, 00:11:29.153 "nvme_admin": false, 00:11:29.153 "nvme_io": false, 00:11:29.153 "nvme_io_md": false, 00:11:29.153 "write_zeroes": true, 00:11:29.153 "zcopy": true, 00:11:29.153 "get_zone_info": false, 00:11:29.153 "zone_management": false, 00:11:29.153 "zone_append": false, 00:11:29.153 "compare": false, 00:11:29.153 "compare_and_write": false, 00:11:29.153 "abort": true, 00:11:29.153 "seek_hole": false, 00:11:29.153 "seek_data": false, 00:11:29.153 "copy": true, 00:11:29.153 "nvme_iov_md": false 00:11:29.153 }, 00:11:29.153 "memory_domains": [ 00:11:29.153 { 00:11:29.153 "dma_device_id": "system", 00:11:29.153 "dma_device_type": 1 00:11:29.153 }, 00:11:29.153 { 00:11:29.153 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:29.153 "dma_device_type": 2 00:11:29.153 } 00:11:29.153 ], 00:11:29.153 "driver_specific": {} 00:11:29.153 } 00:11:29.153 ] 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:29.153 "name": "Existed_Raid", 00:11:29.153 "uuid": "77998354-2506-46ad-9c1c-28acefca2c3d", 00:11:29.153 "strip_size_kb": 64, 00:11:29.153 "state": "configuring", 00:11:29.153 "raid_level": "concat", 00:11:29.153 "superblock": true, 00:11:29.153 "num_base_bdevs": 4, 00:11:29.153 "num_base_bdevs_discovered": 1, 00:11:29.153 "num_base_bdevs_operational": 4, 00:11:29.153 "base_bdevs_list": [ 00:11:29.153 { 00:11:29.153 "name": "BaseBdev1", 00:11:29.153 "uuid": "cae646d7-92a7-42eb-a42e-9f81fb9e9c56", 00:11:29.153 "is_configured": true, 00:11:29.153 "data_offset": 2048, 00:11:29.153 "data_size": 63488 00:11:29.153 }, 00:11:29.153 { 00:11:29.153 "name": "BaseBdev2", 00:11:29.153 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:29.153 "is_configured": false, 00:11:29.153 "data_offset": 0, 00:11:29.153 "data_size": 0 00:11:29.153 }, 00:11:29.153 { 00:11:29.153 "name": "BaseBdev3", 00:11:29.153 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:29.153 "is_configured": false, 00:11:29.153 "data_offset": 0, 00:11:29.153 "data_size": 0 00:11:29.153 }, 00:11:29.153 { 00:11:29.153 "name": "BaseBdev4", 00:11:29.153 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:29.153 "is_configured": false, 00:11:29.153 "data_offset": 0, 00:11:29.153 "data_size": 0 00:11:29.153 } 00:11:29.153 ] 00:11:29.153 }' 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:29.153 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.413 [2024-11-20 15:59:27.634045] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:29.413 [2024-11-20 15:59:27.634196] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.413 [2024-11-20 15:59:27.642118] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:29.413 [2024-11-20 15:59:27.644011] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:29.413 [2024-11-20 15:59:27.644049] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:29.413 [2024-11-20 15:59:27.644059] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:29.413 [2024-11-20 15:59:27.644071] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:29.413 [2024-11-20 15:59:27.644078] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:29.413 [2024-11-20 15:59:27.644088] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.413 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.675 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.675 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:29.675 "name": "Existed_Raid", 00:11:29.675 "uuid": "023a3a3c-8c13-4dfd-ad3b-8e96c2a2d7f9", 00:11:29.675 "strip_size_kb": 64, 00:11:29.675 "state": "configuring", 00:11:29.675 "raid_level": "concat", 00:11:29.675 "superblock": true, 00:11:29.675 "num_base_bdevs": 4, 00:11:29.675 "num_base_bdevs_discovered": 1, 00:11:29.675 "num_base_bdevs_operational": 4, 00:11:29.675 "base_bdevs_list": [ 00:11:29.675 { 00:11:29.675 "name": "BaseBdev1", 00:11:29.675 "uuid": "cae646d7-92a7-42eb-a42e-9f81fb9e9c56", 00:11:29.675 "is_configured": true, 00:11:29.675 "data_offset": 2048, 00:11:29.675 "data_size": 63488 00:11:29.675 }, 00:11:29.675 { 00:11:29.675 "name": "BaseBdev2", 00:11:29.675 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:29.675 "is_configured": false, 00:11:29.675 "data_offset": 0, 00:11:29.675 "data_size": 0 00:11:29.675 }, 00:11:29.675 { 00:11:29.675 "name": "BaseBdev3", 00:11:29.675 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:29.675 "is_configured": false, 00:11:29.675 "data_offset": 0, 00:11:29.675 "data_size": 0 00:11:29.675 }, 00:11:29.675 { 00:11:29.675 "name": "BaseBdev4", 00:11:29.675 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:29.675 "is_configured": false, 00:11:29.675 "data_offset": 0, 00:11:29.675 "data_size": 0 00:11:29.675 } 00:11:29.675 ] 00:11:29.675 }' 00:11:29.675 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:29.675 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.937 15:59:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:29.937 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.937 15:59:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.937 [2024-11-20 15:59:28.009341] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:29.937 BaseBdev2 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.937 [ 00:11:29.937 { 00:11:29.937 "name": "BaseBdev2", 00:11:29.937 "aliases": [ 00:11:29.937 "be2e8660-d05d-46dc-a251-f6f72201641b" 00:11:29.937 ], 00:11:29.937 "product_name": "Malloc disk", 00:11:29.937 "block_size": 512, 00:11:29.937 "num_blocks": 65536, 00:11:29.937 "uuid": "be2e8660-d05d-46dc-a251-f6f72201641b", 00:11:29.937 "assigned_rate_limits": { 00:11:29.937 "rw_ios_per_sec": 0, 00:11:29.937 "rw_mbytes_per_sec": 0, 00:11:29.937 "r_mbytes_per_sec": 0, 00:11:29.937 "w_mbytes_per_sec": 0 00:11:29.937 }, 00:11:29.937 "claimed": true, 00:11:29.937 "claim_type": "exclusive_write", 00:11:29.937 "zoned": false, 00:11:29.937 "supported_io_types": { 00:11:29.937 "read": true, 00:11:29.937 "write": true, 00:11:29.937 "unmap": true, 00:11:29.937 "flush": true, 00:11:29.937 "reset": true, 00:11:29.937 "nvme_admin": false, 00:11:29.937 "nvme_io": false, 00:11:29.937 "nvme_io_md": false, 00:11:29.937 "write_zeroes": true, 00:11:29.937 "zcopy": true, 00:11:29.937 "get_zone_info": false, 00:11:29.937 "zone_management": false, 00:11:29.937 "zone_append": false, 00:11:29.937 "compare": false, 00:11:29.937 "compare_and_write": false, 00:11:29.937 "abort": true, 00:11:29.937 "seek_hole": false, 00:11:29.937 "seek_data": false, 00:11:29.937 "copy": true, 00:11:29.937 "nvme_iov_md": false 00:11:29.937 }, 00:11:29.937 "memory_domains": [ 00:11:29.937 { 00:11:29.937 "dma_device_id": "system", 00:11:29.937 "dma_device_type": 1 00:11:29.937 }, 00:11:29.937 { 00:11:29.937 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:29.937 "dma_device_type": 2 00:11:29.937 } 00:11:29.937 ], 00:11:29.937 "driver_specific": {} 00:11:29.937 } 00:11:29.937 ] 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:29.937 "name": "Existed_Raid", 00:11:29.937 "uuid": "023a3a3c-8c13-4dfd-ad3b-8e96c2a2d7f9", 00:11:29.937 "strip_size_kb": 64, 00:11:29.937 "state": "configuring", 00:11:29.937 "raid_level": "concat", 00:11:29.937 "superblock": true, 00:11:29.937 "num_base_bdevs": 4, 00:11:29.937 "num_base_bdevs_discovered": 2, 00:11:29.937 "num_base_bdevs_operational": 4, 00:11:29.937 "base_bdevs_list": [ 00:11:29.937 { 00:11:29.937 "name": "BaseBdev1", 00:11:29.937 "uuid": "cae646d7-92a7-42eb-a42e-9f81fb9e9c56", 00:11:29.937 "is_configured": true, 00:11:29.937 "data_offset": 2048, 00:11:29.937 "data_size": 63488 00:11:29.937 }, 00:11:29.937 { 00:11:29.937 "name": "BaseBdev2", 00:11:29.937 "uuid": "be2e8660-d05d-46dc-a251-f6f72201641b", 00:11:29.937 "is_configured": true, 00:11:29.937 "data_offset": 2048, 00:11:29.937 "data_size": 63488 00:11:29.937 }, 00:11:29.937 { 00:11:29.937 "name": "BaseBdev3", 00:11:29.937 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:29.937 "is_configured": false, 00:11:29.937 "data_offset": 0, 00:11:29.937 "data_size": 0 00:11:29.937 }, 00:11:29.937 { 00:11:29.937 "name": "BaseBdev4", 00:11:29.937 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:29.937 "is_configured": false, 00:11:29.937 "data_offset": 0, 00:11:29.937 "data_size": 0 00:11:29.937 } 00:11:29.937 ] 00:11:29.937 }' 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:29.937 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.199 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:30.199 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.199 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.199 [2024-11-20 15:59:28.404641] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:30.199 BaseBdev3 00:11:30.199 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.199 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:11:30.199 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:11:30.199 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:30.199 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:30.199 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:30.199 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:30.199 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:30.199 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.199 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.199 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.199 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:30.199 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.199 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.199 [ 00:11:30.199 { 00:11:30.199 "name": "BaseBdev3", 00:11:30.199 "aliases": [ 00:11:30.199 "932fbeb1-b9b7-4465-ac53-ce2fc5ac3926" 00:11:30.199 ], 00:11:30.199 "product_name": "Malloc disk", 00:11:30.199 "block_size": 512, 00:11:30.199 "num_blocks": 65536, 00:11:30.199 "uuid": "932fbeb1-b9b7-4465-ac53-ce2fc5ac3926", 00:11:30.199 "assigned_rate_limits": { 00:11:30.199 "rw_ios_per_sec": 0, 00:11:30.199 "rw_mbytes_per_sec": 0, 00:11:30.199 "r_mbytes_per_sec": 0, 00:11:30.199 "w_mbytes_per_sec": 0 00:11:30.199 }, 00:11:30.199 "claimed": true, 00:11:30.199 "claim_type": "exclusive_write", 00:11:30.199 "zoned": false, 00:11:30.199 "supported_io_types": { 00:11:30.199 "read": true, 00:11:30.199 "write": true, 00:11:30.199 "unmap": true, 00:11:30.199 "flush": true, 00:11:30.199 "reset": true, 00:11:30.199 "nvme_admin": false, 00:11:30.199 "nvme_io": false, 00:11:30.199 "nvme_io_md": false, 00:11:30.199 "write_zeroes": true, 00:11:30.199 "zcopy": true, 00:11:30.199 "get_zone_info": false, 00:11:30.199 "zone_management": false, 00:11:30.199 "zone_append": false, 00:11:30.199 "compare": false, 00:11:30.199 "compare_and_write": false, 00:11:30.199 "abort": true, 00:11:30.200 "seek_hole": false, 00:11:30.200 "seek_data": false, 00:11:30.200 "copy": true, 00:11:30.200 "nvme_iov_md": false 00:11:30.200 }, 00:11:30.200 "memory_domains": [ 00:11:30.200 { 00:11:30.200 "dma_device_id": "system", 00:11:30.200 "dma_device_type": 1 00:11:30.200 }, 00:11:30.200 { 00:11:30.200 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:30.200 "dma_device_type": 2 00:11:30.200 } 00:11:30.200 ], 00:11:30.200 "driver_specific": {} 00:11:30.200 } 00:11:30.200 ] 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.200 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.460 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.460 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:30.460 "name": "Existed_Raid", 00:11:30.460 "uuid": "023a3a3c-8c13-4dfd-ad3b-8e96c2a2d7f9", 00:11:30.460 "strip_size_kb": 64, 00:11:30.460 "state": "configuring", 00:11:30.460 "raid_level": "concat", 00:11:30.460 "superblock": true, 00:11:30.460 "num_base_bdevs": 4, 00:11:30.460 "num_base_bdevs_discovered": 3, 00:11:30.460 "num_base_bdevs_operational": 4, 00:11:30.460 "base_bdevs_list": [ 00:11:30.460 { 00:11:30.460 "name": "BaseBdev1", 00:11:30.460 "uuid": "cae646d7-92a7-42eb-a42e-9f81fb9e9c56", 00:11:30.460 "is_configured": true, 00:11:30.460 "data_offset": 2048, 00:11:30.460 "data_size": 63488 00:11:30.460 }, 00:11:30.460 { 00:11:30.460 "name": "BaseBdev2", 00:11:30.460 "uuid": "be2e8660-d05d-46dc-a251-f6f72201641b", 00:11:30.460 "is_configured": true, 00:11:30.460 "data_offset": 2048, 00:11:30.460 "data_size": 63488 00:11:30.460 }, 00:11:30.460 { 00:11:30.460 "name": "BaseBdev3", 00:11:30.460 "uuid": "932fbeb1-b9b7-4465-ac53-ce2fc5ac3926", 00:11:30.460 "is_configured": true, 00:11:30.460 "data_offset": 2048, 00:11:30.460 "data_size": 63488 00:11:30.460 }, 00:11:30.460 { 00:11:30.460 "name": "BaseBdev4", 00:11:30.460 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:30.460 "is_configured": false, 00:11:30.460 "data_offset": 0, 00:11:30.460 "data_size": 0 00:11:30.460 } 00:11:30.460 ] 00:11:30.460 }' 00:11:30.460 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:30.460 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.723 [2024-11-20 15:59:28.795256] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:30.723 [2024-11-20 15:59:28.795477] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:11:30.723 [2024-11-20 15:59:28.795489] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:30.723 [2024-11-20 15:59:28.795773] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:11:30.723 [2024-11-20 15:59:28.795906] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:11:30.723 [2024-11-20 15:59:28.795926] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:11:30.723 BaseBdev4 00:11:30.723 [2024-11-20 15:59:28.796052] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.723 [ 00:11:30.723 { 00:11:30.723 "name": "BaseBdev4", 00:11:30.723 "aliases": [ 00:11:30.723 "58a560be-0870-49d5-8d82-be6996fe7c92" 00:11:30.723 ], 00:11:30.723 "product_name": "Malloc disk", 00:11:30.723 "block_size": 512, 00:11:30.723 "num_blocks": 65536, 00:11:30.723 "uuid": "58a560be-0870-49d5-8d82-be6996fe7c92", 00:11:30.723 "assigned_rate_limits": { 00:11:30.723 "rw_ios_per_sec": 0, 00:11:30.723 "rw_mbytes_per_sec": 0, 00:11:30.723 "r_mbytes_per_sec": 0, 00:11:30.723 "w_mbytes_per_sec": 0 00:11:30.723 }, 00:11:30.723 "claimed": true, 00:11:30.723 "claim_type": "exclusive_write", 00:11:30.723 "zoned": false, 00:11:30.723 "supported_io_types": { 00:11:30.723 "read": true, 00:11:30.723 "write": true, 00:11:30.723 "unmap": true, 00:11:30.723 "flush": true, 00:11:30.723 "reset": true, 00:11:30.723 "nvme_admin": false, 00:11:30.723 "nvme_io": false, 00:11:30.723 "nvme_io_md": false, 00:11:30.723 "write_zeroes": true, 00:11:30.723 "zcopy": true, 00:11:30.723 "get_zone_info": false, 00:11:30.723 "zone_management": false, 00:11:30.723 "zone_append": false, 00:11:30.723 "compare": false, 00:11:30.723 "compare_and_write": false, 00:11:30.723 "abort": true, 00:11:30.723 "seek_hole": false, 00:11:30.723 "seek_data": false, 00:11:30.723 "copy": true, 00:11:30.723 "nvme_iov_md": false 00:11:30.723 }, 00:11:30.723 "memory_domains": [ 00:11:30.723 { 00:11:30.723 "dma_device_id": "system", 00:11:30.723 "dma_device_type": 1 00:11:30.723 }, 00:11:30.723 { 00:11:30.723 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:30.723 "dma_device_type": 2 00:11:30.723 } 00:11:30.723 ], 00:11:30.723 "driver_specific": {} 00:11:30.723 } 00:11:30.723 ] 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.723 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:30.723 "name": "Existed_Raid", 00:11:30.723 "uuid": "023a3a3c-8c13-4dfd-ad3b-8e96c2a2d7f9", 00:11:30.723 "strip_size_kb": 64, 00:11:30.723 "state": "online", 00:11:30.723 "raid_level": "concat", 00:11:30.723 "superblock": true, 00:11:30.723 "num_base_bdevs": 4, 00:11:30.723 "num_base_bdevs_discovered": 4, 00:11:30.723 "num_base_bdevs_operational": 4, 00:11:30.723 "base_bdevs_list": [ 00:11:30.723 { 00:11:30.723 "name": "BaseBdev1", 00:11:30.723 "uuid": "cae646d7-92a7-42eb-a42e-9f81fb9e9c56", 00:11:30.723 "is_configured": true, 00:11:30.723 "data_offset": 2048, 00:11:30.723 "data_size": 63488 00:11:30.723 }, 00:11:30.723 { 00:11:30.723 "name": "BaseBdev2", 00:11:30.723 "uuid": "be2e8660-d05d-46dc-a251-f6f72201641b", 00:11:30.723 "is_configured": true, 00:11:30.723 "data_offset": 2048, 00:11:30.723 "data_size": 63488 00:11:30.723 }, 00:11:30.723 { 00:11:30.723 "name": "BaseBdev3", 00:11:30.723 "uuid": "932fbeb1-b9b7-4465-ac53-ce2fc5ac3926", 00:11:30.723 "is_configured": true, 00:11:30.723 "data_offset": 2048, 00:11:30.723 "data_size": 63488 00:11:30.723 }, 00:11:30.723 { 00:11:30.723 "name": "BaseBdev4", 00:11:30.724 "uuid": "58a560be-0870-49d5-8d82-be6996fe7c92", 00:11:30.724 "is_configured": true, 00:11:30.724 "data_offset": 2048, 00:11:30.724 "data_size": 63488 00:11:30.724 } 00:11:30.724 ] 00:11:30.724 }' 00:11:30.724 15:59:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:30.724 15:59:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.028 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:11:31.028 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:31.029 [2024-11-20 15:59:29.147785] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:31.029 "name": "Existed_Raid", 00:11:31.029 "aliases": [ 00:11:31.029 "023a3a3c-8c13-4dfd-ad3b-8e96c2a2d7f9" 00:11:31.029 ], 00:11:31.029 "product_name": "Raid Volume", 00:11:31.029 "block_size": 512, 00:11:31.029 "num_blocks": 253952, 00:11:31.029 "uuid": "023a3a3c-8c13-4dfd-ad3b-8e96c2a2d7f9", 00:11:31.029 "assigned_rate_limits": { 00:11:31.029 "rw_ios_per_sec": 0, 00:11:31.029 "rw_mbytes_per_sec": 0, 00:11:31.029 "r_mbytes_per_sec": 0, 00:11:31.029 "w_mbytes_per_sec": 0 00:11:31.029 }, 00:11:31.029 "claimed": false, 00:11:31.029 "zoned": false, 00:11:31.029 "supported_io_types": { 00:11:31.029 "read": true, 00:11:31.029 "write": true, 00:11:31.029 "unmap": true, 00:11:31.029 "flush": true, 00:11:31.029 "reset": true, 00:11:31.029 "nvme_admin": false, 00:11:31.029 "nvme_io": false, 00:11:31.029 "nvme_io_md": false, 00:11:31.029 "write_zeroes": true, 00:11:31.029 "zcopy": false, 00:11:31.029 "get_zone_info": false, 00:11:31.029 "zone_management": false, 00:11:31.029 "zone_append": false, 00:11:31.029 "compare": false, 00:11:31.029 "compare_and_write": false, 00:11:31.029 "abort": false, 00:11:31.029 "seek_hole": false, 00:11:31.029 "seek_data": false, 00:11:31.029 "copy": false, 00:11:31.029 "nvme_iov_md": false 00:11:31.029 }, 00:11:31.029 "memory_domains": [ 00:11:31.029 { 00:11:31.029 "dma_device_id": "system", 00:11:31.029 "dma_device_type": 1 00:11:31.029 }, 00:11:31.029 { 00:11:31.029 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:31.029 "dma_device_type": 2 00:11:31.029 }, 00:11:31.029 { 00:11:31.029 "dma_device_id": "system", 00:11:31.029 "dma_device_type": 1 00:11:31.029 }, 00:11:31.029 { 00:11:31.029 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:31.029 "dma_device_type": 2 00:11:31.029 }, 00:11:31.029 { 00:11:31.029 "dma_device_id": "system", 00:11:31.029 "dma_device_type": 1 00:11:31.029 }, 00:11:31.029 { 00:11:31.029 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:31.029 "dma_device_type": 2 00:11:31.029 }, 00:11:31.029 { 00:11:31.029 "dma_device_id": "system", 00:11:31.029 "dma_device_type": 1 00:11:31.029 }, 00:11:31.029 { 00:11:31.029 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:31.029 "dma_device_type": 2 00:11:31.029 } 00:11:31.029 ], 00:11:31.029 "driver_specific": { 00:11:31.029 "raid": { 00:11:31.029 "uuid": "023a3a3c-8c13-4dfd-ad3b-8e96c2a2d7f9", 00:11:31.029 "strip_size_kb": 64, 00:11:31.029 "state": "online", 00:11:31.029 "raid_level": "concat", 00:11:31.029 "superblock": true, 00:11:31.029 "num_base_bdevs": 4, 00:11:31.029 "num_base_bdevs_discovered": 4, 00:11:31.029 "num_base_bdevs_operational": 4, 00:11:31.029 "base_bdevs_list": [ 00:11:31.029 { 00:11:31.029 "name": "BaseBdev1", 00:11:31.029 "uuid": "cae646d7-92a7-42eb-a42e-9f81fb9e9c56", 00:11:31.029 "is_configured": true, 00:11:31.029 "data_offset": 2048, 00:11:31.029 "data_size": 63488 00:11:31.029 }, 00:11:31.029 { 00:11:31.029 "name": "BaseBdev2", 00:11:31.029 "uuid": "be2e8660-d05d-46dc-a251-f6f72201641b", 00:11:31.029 "is_configured": true, 00:11:31.029 "data_offset": 2048, 00:11:31.029 "data_size": 63488 00:11:31.029 }, 00:11:31.029 { 00:11:31.029 "name": "BaseBdev3", 00:11:31.029 "uuid": "932fbeb1-b9b7-4465-ac53-ce2fc5ac3926", 00:11:31.029 "is_configured": true, 00:11:31.029 "data_offset": 2048, 00:11:31.029 "data_size": 63488 00:11:31.029 }, 00:11:31.029 { 00:11:31.029 "name": "BaseBdev4", 00:11:31.029 "uuid": "58a560be-0870-49d5-8d82-be6996fe7c92", 00:11:31.029 "is_configured": true, 00:11:31.029 "data_offset": 2048, 00:11:31.029 "data_size": 63488 00:11:31.029 } 00:11:31.029 ] 00:11:31.029 } 00:11:31.029 } 00:11:31.029 }' 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:11:31.029 BaseBdev2 00:11:31.029 BaseBdev3 00:11:31.029 BaseBdev4' 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.029 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.292 [2024-11-20 15:59:29.387518] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:31.292 [2024-11-20 15:59:29.387646] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:31.292 [2024-11-20 15:59:29.387726] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:31.292 "name": "Existed_Raid", 00:11:31.292 "uuid": "023a3a3c-8c13-4dfd-ad3b-8e96c2a2d7f9", 00:11:31.292 "strip_size_kb": 64, 00:11:31.292 "state": "offline", 00:11:31.292 "raid_level": "concat", 00:11:31.292 "superblock": true, 00:11:31.292 "num_base_bdevs": 4, 00:11:31.292 "num_base_bdevs_discovered": 3, 00:11:31.292 "num_base_bdevs_operational": 3, 00:11:31.292 "base_bdevs_list": [ 00:11:31.292 { 00:11:31.292 "name": null, 00:11:31.292 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:31.292 "is_configured": false, 00:11:31.292 "data_offset": 0, 00:11:31.292 "data_size": 63488 00:11:31.292 }, 00:11:31.292 { 00:11:31.292 "name": "BaseBdev2", 00:11:31.292 "uuid": "be2e8660-d05d-46dc-a251-f6f72201641b", 00:11:31.292 "is_configured": true, 00:11:31.292 "data_offset": 2048, 00:11:31.292 "data_size": 63488 00:11:31.292 }, 00:11:31.292 { 00:11:31.292 "name": "BaseBdev3", 00:11:31.292 "uuid": "932fbeb1-b9b7-4465-ac53-ce2fc5ac3926", 00:11:31.292 "is_configured": true, 00:11:31.292 "data_offset": 2048, 00:11:31.292 "data_size": 63488 00:11:31.292 }, 00:11:31.292 { 00:11:31.292 "name": "BaseBdev4", 00:11:31.292 "uuid": "58a560be-0870-49d5-8d82-be6996fe7c92", 00:11:31.292 "is_configured": true, 00:11:31.292 "data_offset": 2048, 00:11:31.292 "data_size": 63488 00:11:31.292 } 00:11:31.292 ] 00:11:31.292 }' 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:31.292 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.554 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:11:31.554 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:31.554 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:31.554 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:31.554 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.554 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.554 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.554 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:31.554 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:31.554 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:11:31.554 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.554 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.554 [2024-11-20 15:59:29.786657] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.815 [2024-11-20 15:59:29.885018] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.815 15:59:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.815 [2024-11-20 15:59:29.984032] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:11:31.815 [2024-11-20 15:59:29.984075] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:11:31.815 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.815 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:31.815 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:31.815 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:11:31.815 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:31.815 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.815 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.815 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.077 BaseBdev2 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.077 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.077 [ 00:11:32.077 { 00:11:32.077 "name": "BaseBdev2", 00:11:32.077 "aliases": [ 00:11:32.077 "2c5d827c-a20e-4ad8-b5c2-2ea145391ad4" 00:11:32.077 ], 00:11:32.077 "product_name": "Malloc disk", 00:11:32.077 "block_size": 512, 00:11:32.077 "num_blocks": 65536, 00:11:32.077 "uuid": "2c5d827c-a20e-4ad8-b5c2-2ea145391ad4", 00:11:32.077 "assigned_rate_limits": { 00:11:32.077 "rw_ios_per_sec": 0, 00:11:32.077 "rw_mbytes_per_sec": 0, 00:11:32.077 "r_mbytes_per_sec": 0, 00:11:32.077 "w_mbytes_per_sec": 0 00:11:32.077 }, 00:11:32.077 "claimed": false, 00:11:32.077 "zoned": false, 00:11:32.077 "supported_io_types": { 00:11:32.077 "read": true, 00:11:32.077 "write": true, 00:11:32.077 "unmap": true, 00:11:32.077 "flush": true, 00:11:32.077 "reset": true, 00:11:32.078 "nvme_admin": false, 00:11:32.078 "nvme_io": false, 00:11:32.078 "nvme_io_md": false, 00:11:32.078 "write_zeroes": true, 00:11:32.078 "zcopy": true, 00:11:32.078 "get_zone_info": false, 00:11:32.078 "zone_management": false, 00:11:32.078 "zone_append": false, 00:11:32.078 "compare": false, 00:11:32.078 "compare_and_write": false, 00:11:32.078 "abort": true, 00:11:32.078 "seek_hole": false, 00:11:32.078 "seek_data": false, 00:11:32.078 "copy": true, 00:11:32.078 "nvme_iov_md": false 00:11:32.078 }, 00:11:32.078 "memory_domains": [ 00:11:32.078 { 00:11:32.078 "dma_device_id": "system", 00:11:32.078 "dma_device_type": 1 00:11:32.078 }, 00:11:32.078 { 00:11:32.078 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:32.078 "dma_device_type": 2 00:11:32.078 } 00:11:32.078 ], 00:11:32.078 "driver_specific": {} 00:11:32.078 } 00:11:32.078 ] 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.078 BaseBdev3 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.078 [ 00:11:32.078 { 00:11:32.078 "name": "BaseBdev3", 00:11:32.078 "aliases": [ 00:11:32.078 "a0f2d327-f572-4c13-8054-9852c394411a" 00:11:32.078 ], 00:11:32.078 "product_name": "Malloc disk", 00:11:32.078 "block_size": 512, 00:11:32.078 "num_blocks": 65536, 00:11:32.078 "uuid": "a0f2d327-f572-4c13-8054-9852c394411a", 00:11:32.078 "assigned_rate_limits": { 00:11:32.078 "rw_ios_per_sec": 0, 00:11:32.078 "rw_mbytes_per_sec": 0, 00:11:32.078 "r_mbytes_per_sec": 0, 00:11:32.078 "w_mbytes_per_sec": 0 00:11:32.078 }, 00:11:32.078 "claimed": false, 00:11:32.078 "zoned": false, 00:11:32.078 "supported_io_types": { 00:11:32.078 "read": true, 00:11:32.078 "write": true, 00:11:32.078 "unmap": true, 00:11:32.078 "flush": true, 00:11:32.078 "reset": true, 00:11:32.078 "nvme_admin": false, 00:11:32.078 "nvme_io": false, 00:11:32.078 "nvme_io_md": false, 00:11:32.078 "write_zeroes": true, 00:11:32.078 "zcopy": true, 00:11:32.078 "get_zone_info": false, 00:11:32.078 "zone_management": false, 00:11:32.078 "zone_append": false, 00:11:32.078 "compare": false, 00:11:32.078 "compare_and_write": false, 00:11:32.078 "abort": true, 00:11:32.078 "seek_hole": false, 00:11:32.078 "seek_data": false, 00:11:32.078 "copy": true, 00:11:32.078 "nvme_iov_md": false 00:11:32.078 }, 00:11:32.078 "memory_domains": [ 00:11:32.078 { 00:11:32.078 "dma_device_id": "system", 00:11:32.078 "dma_device_type": 1 00:11:32.078 }, 00:11:32.078 { 00:11:32.078 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:32.078 "dma_device_type": 2 00:11:32.078 } 00:11:32.078 ], 00:11:32.078 "driver_specific": {} 00:11:32.078 } 00:11:32.078 ] 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.078 BaseBdev4 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.078 [ 00:11:32.078 { 00:11:32.078 "name": "BaseBdev4", 00:11:32.078 "aliases": [ 00:11:32.078 "9fa36fa8-1027-45fe-97e9-5734edd38091" 00:11:32.078 ], 00:11:32.078 "product_name": "Malloc disk", 00:11:32.078 "block_size": 512, 00:11:32.078 "num_blocks": 65536, 00:11:32.078 "uuid": "9fa36fa8-1027-45fe-97e9-5734edd38091", 00:11:32.078 "assigned_rate_limits": { 00:11:32.078 "rw_ios_per_sec": 0, 00:11:32.078 "rw_mbytes_per_sec": 0, 00:11:32.078 "r_mbytes_per_sec": 0, 00:11:32.078 "w_mbytes_per_sec": 0 00:11:32.078 }, 00:11:32.078 "claimed": false, 00:11:32.078 "zoned": false, 00:11:32.078 "supported_io_types": { 00:11:32.078 "read": true, 00:11:32.078 "write": true, 00:11:32.078 "unmap": true, 00:11:32.078 "flush": true, 00:11:32.078 "reset": true, 00:11:32.078 "nvme_admin": false, 00:11:32.078 "nvme_io": false, 00:11:32.078 "nvme_io_md": false, 00:11:32.078 "write_zeroes": true, 00:11:32.078 "zcopy": true, 00:11:32.078 "get_zone_info": false, 00:11:32.078 "zone_management": false, 00:11:32.078 "zone_append": false, 00:11:32.078 "compare": false, 00:11:32.078 "compare_and_write": false, 00:11:32.078 "abort": true, 00:11:32.078 "seek_hole": false, 00:11:32.078 "seek_data": false, 00:11:32.078 "copy": true, 00:11:32.078 "nvme_iov_md": false 00:11:32.078 }, 00:11:32.078 "memory_domains": [ 00:11:32.078 { 00:11:32.078 "dma_device_id": "system", 00:11:32.078 "dma_device_type": 1 00:11:32.078 }, 00:11:32.078 { 00:11:32.078 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:32.078 "dma_device_type": 2 00:11:32.078 } 00:11:32.078 ], 00:11:32.078 "driver_specific": {} 00:11:32.078 } 00:11:32.078 ] 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.078 [2024-11-20 15:59:30.252650] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:32.078 [2024-11-20 15:59:30.252795] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:32.078 [2024-11-20 15:59:30.252866] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:32.078 [2024-11-20 15:59:30.254752] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:32.078 [2024-11-20 15:59:30.254879] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:32.078 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:32.079 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:32.079 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:32.079 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:32.079 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:32.079 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:32.079 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:32.079 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:32.079 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:32.079 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:32.079 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:32.079 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.079 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.079 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.079 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:32.079 "name": "Existed_Raid", 00:11:32.079 "uuid": "a92153ed-1a99-43d7-a82d-0e0dd960cdd3", 00:11:32.079 "strip_size_kb": 64, 00:11:32.079 "state": "configuring", 00:11:32.079 "raid_level": "concat", 00:11:32.079 "superblock": true, 00:11:32.079 "num_base_bdevs": 4, 00:11:32.079 "num_base_bdevs_discovered": 3, 00:11:32.079 "num_base_bdevs_operational": 4, 00:11:32.079 "base_bdevs_list": [ 00:11:32.079 { 00:11:32.079 "name": "BaseBdev1", 00:11:32.079 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:32.079 "is_configured": false, 00:11:32.079 "data_offset": 0, 00:11:32.079 "data_size": 0 00:11:32.079 }, 00:11:32.079 { 00:11:32.079 "name": "BaseBdev2", 00:11:32.079 "uuid": "2c5d827c-a20e-4ad8-b5c2-2ea145391ad4", 00:11:32.079 "is_configured": true, 00:11:32.079 "data_offset": 2048, 00:11:32.079 "data_size": 63488 00:11:32.079 }, 00:11:32.079 { 00:11:32.079 "name": "BaseBdev3", 00:11:32.079 "uuid": "a0f2d327-f572-4c13-8054-9852c394411a", 00:11:32.079 "is_configured": true, 00:11:32.079 "data_offset": 2048, 00:11:32.079 "data_size": 63488 00:11:32.079 }, 00:11:32.079 { 00:11:32.079 "name": "BaseBdev4", 00:11:32.079 "uuid": "9fa36fa8-1027-45fe-97e9-5734edd38091", 00:11:32.079 "is_configured": true, 00:11:32.079 "data_offset": 2048, 00:11:32.079 "data_size": 63488 00:11:32.079 } 00:11:32.079 ] 00:11:32.079 }' 00:11:32.079 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:32.079 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.338 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:11:32.338 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.338 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.597 [2024-11-20 15:59:30.588756] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:32.597 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.597 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:32.597 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:32.597 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:32.597 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:32.597 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:32.597 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:32.597 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:32.597 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:32.597 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:32.597 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:32.597 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:32.597 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.597 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.597 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:32.597 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.597 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:32.597 "name": "Existed_Raid", 00:11:32.597 "uuid": "a92153ed-1a99-43d7-a82d-0e0dd960cdd3", 00:11:32.597 "strip_size_kb": 64, 00:11:32.597 "state": "configuring", 00:11:32.597 "raid_level": "concat", 00:11:32.597 "superblock": true, 00:11:32.597 "num_base_bdevs": 4, 00:11:32.597 "num_base_bdevs_discovered": 2, 00:11:32.597 "num_base_bdevs_operational": 4, 00:11:32.597 "base_bdevs_list": [ 00:11:32.597 { 00:11:32.597 "name": "BaseBdev1", 00:11:32.597 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:32.597 "is_configured": false, 00:11:32.597 "data_offset": 0, 00:11:32.597 "data_size": 0 00:11:32.597 }, 00:11:32.597 { 00:11:32.597 "name": null, 00:11:32.597 "uuid": "2c5d827c-a20e-4ad8-b5c2-2ea145391ad4", 00:11:32.597 "is_configured": false, 00:11:32.597 "data_offset": 0, 00:11:32.597 "data_size": 63488 00:11:32.597 }, 00:11:32.597 { 00:11:32.597 "name": "BaseBdev3", 00:11:32.597 "uuid": "a0f2d327-f572-4c13-8054-9852c394411a", 00:11:32.597 "is_configured": true, 00:11:32.597 "data_offset": 2048, 00:11:32.597 "data_size": 63488 00:11:32.598 }, 00:11:32.598 { 00:11:32.598 "name": "BaseBdev4", 00:11:32.598 "uuid": "9fa36fa8-1027-45fe-97e9-5734edd38091", 00:11:32.598 "is_configured": true, 00:11:32.598 "data_offset": 2048, 00:11:32.598 "data_size": 63488 00:11:32.598 } 00:11:32.598 ] 00:11:32.598 }' 00:11:32.598 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:32.598 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.858 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:32.858 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:32.858 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.858 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.858 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.859 [2024-11-20 15:59:30.957030] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:32.859 BaseBdev1 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.859 [ 00:11:32.859 { 00:11:32.859 "name": "BaseBdev1", 00:11:32.859 "aliases": [ 00:11:32.859 "730300c5-2cfd-470f-afbb-f4d249862836" 00:11:32.859 ], 00:11:32.859 "product_name": "Malloc disk", 00:11:32.859 "block_size": 512, 00:11:32.859 "num_blocks": 65536, 00:11:32.859 "uuid": "730300c5-2cfd-470f-afbb-f4d249862836", 00:11:32.859 "assigned_rate_limits": { 00:11:32.859 "rw_ios_per_sec": 0, 00:11:32.859 "rw_mbytes_per_sec": 0, 00:11:32.859 "r_mbytes_per_sec": 0, 00:11:32.859 "w_mbytes_per_sec": 0 00:11:32.859 }, 00:11:32.859 "claimed": true, 00:11:32.859 "claim_type": "exclusive_write", 00:11:32.859 "zoned": false, 00:11:32.859 "supported_io_types": { 00:11:32.859 "read": true, 00:11:32.859 "write": true, 00:11:32.859 "unmap": true, 00:11:32.859 "flush": true, 00:11:32.859 "reset": true, 00:11:32.859 "nvme_admin": false, 00:11:32.859 "nvme_io": false, 00:11:32.859 "nvme_io_md": false, 00:11:32.859 "write_zeroes": true, 00:11:32.859 "zcopy": true, 00:11:32.859 "get_zone_info": false, 00:11:32.859 "zone_management": false, 00:11:32.859 "zone_append": false, 00:11:32.859 "compare": false, 00:11:32.859 "compare_and_write": false, 00:11:32.859 "abort": true, 00:11:32.859 "seek_hole": false, 00:11:32.859 "seek_data": false, 00:11:32.859 "copy": true, 00:11:32.859 "nvme_iov_md": false 00:11:32.859 }, 00:11:32.859 "memory_domains": [ 00:11:32.859 { 00:11:32.859 "dma_device_id": "system", 00:11:32.859 "dma_device_type": 1 00:11:32.859 }, 00:11:32.859 { 00:11:32.859 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:32.859 "dma_device_type": 2 00:11:32.859 } 00:11:32.859 ], 00:11:32.859 "driver_specific": {} 00:11:32.859 } 00:11:32.859 ] 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.859 15:59:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:32.859 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.859 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:32.859 "name": "Existed_Raid", 00:11:32.859 "uuid": "a92153ed-1a99-43d7-a82d-0e0dd960cdd3", 00:11:32.859 "strip_size_kb": 64, 00:11:32.859 "state": "configuring", 00:11:32.859 "raid_level": "concat", 00:11:32.859 "superblock": true, 00:11:32.859 "num_base_bdevs": 4, 00:11:32.859 "num_base_bdevs_discovered": 3, 00:11:32.859 "num_base_bdevs_operational": 4, 00:11:32.859 "base_bdevs_list": [ 00:11:32.859 { 00:11:32.859 "name": "BaseBdev1", 00:11:32.859 "uuid": "730300c5-2cfd-470f-afbb-f4d249862836", 00:11:32.859 "is_configured": true, 00:11:32.859 "data_offset": 2048, 00:11:32.859 "data_size": 63488 00:11:32.859 }, 00:11:32.859 { 00:11:32.859 "name": null, 00:11:32.859 "uuid": "2c5d827c-a20e-4ad8-b5c2-2ea145391ad4", 00:11:32.859 "is_configured": false, 00:11:32.859 "data_offset": 0, 00:11:32.859 "data_size": 63488 00:11:32.859 }, 00:11:32.859 { 00:11:32.859 "name": "BaseBdev3", 00:11:32.859 "uuid": "a0f2d327-f572-4c13-8054-9852c394411a", 00:11:32.859 "is_configured": true, 00:11:32.859 "data_offset": 2048, 00:11:32.859 "data_size": 63488 00:11:32.859 }, 00:11:32.859 { 00:11:32.859 "name": "BaseBdev4", 00:11:32.859 "uuid": "9fa36fa8-1027-45fe-97e9-5734edd38091", 00:11:32.859 "is_configured": true, 00:11:32.859 "data_offset": 2048, 00:11:32.859 "data_size": 63488 00:11:32.859 } 00:11:32.859 ] 00:11:32.859 }' 00:11:32.859 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:32.859 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.120 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:33.120 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.120 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.120 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.120 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.120 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:11:33.120 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:11:33.120 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.120 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.120 [2024-11-20 15:59:31.353207] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:33.120 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.120 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:33.120 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:33.120 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:33.120 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:33.120 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:33.121 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:33.121 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:33.121 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:33.121 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:33.121 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:33.121 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.121 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.121 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.121 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:33.380 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.380 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:33.380 "name": "Existed_Raid", 00:11:33.380 "uuid": "a92153ed-1a99-43d7-a82d-0e0dd960cdd3", 00:11:33.380 "strip_size_kb": 64, 00:11:33.380 "state": "configuring", 00:11:33.380 "raid_level": "concat", 00:11:33.380 "superblock": true, 00:11:33.380 "num_base_bdevs": 4, 00:11:33.380 "num_base_bdevs_discovered": 2, 00:11:33.380 "num_base_bdevs_operational": 4, 00:11:33.380 "base_bdevs_list": [ 00:11:33.380 { 00:11:33.380 "name": "BaseBdev1", 00:11:33.380 "uuid": "730300c5-2cfd-470f-afbb-f4d249862836", 00:11:33.380 "is_configured": true, 00:11:33.380 "data_offset": 2048, 00:11:33.380 "data_size": 63488 00:11:33.380 }, 00:11:33.380 { 00:11:33.380 "name": null, 00:11:33.380 "uuid": "2c5d827c-a20e-4ad8-b5c2-2ea145391ad4", 00:11:33.380 "is_configured": false, 00:11:33.380 "data_offset": 0, 00:11:33.380 "data_size": 63488 00:11:33.380 }, 00:11:33.380 { 00:11:33.380 "name": null, 00:11:33.380 "uuid": "a0f2d327-f572-4c13-8054-9852c394411a", 00:11:33.380 "is_configured": false, 00:11:33.380 "data_offset": 0, 00:11:33.380 "data_size": 63488 00:11:33.380 }, 00:11:33.380 { 00:11:33.380 "name": "BaseBdev4", 00:11:33.380 "uuid": "9fa36fa8-1027-45fe-97e9-5734edd38091", 00:11:33.380 "is_configured": true, 00:11:33.380 "data_offset": 2048, 00:11:33.380 "data_size": 63488 00:11:33.380 } 00:11:33.380 ] 00:11:33.380 }' 00:11:33.380 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:33.380 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.639 [2024-11-20 15:59:31.725264] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:33.639 "name": "Existed_Raid", 00:11:33.639 "uuid": "a92153ed-1a99-43d7-a82d-0e0dd960cdd3", 00:11:33.639 "strip_size_kb": 64, 00:11:33.639 "state": "configuring", 00:11:33.639 "raid_level": "concat", 00:11:33.639 "superblock": true, 00:11:33.639 "num_base_bdevs": 4, 00:11:33.639 "num_base_bdevs_discovered": 3, 00:11:33.639 "num_base_bdevs_operational": 4, 00:11:33.639 "base_bdevs_list": [ 00:11:33.639 { 00:11:33.639 "name": "BaseBdev1", 00:11:33.639 "uuid": "730300c5-2cfd-470f-afbb-f4d249862836", 00:11:33.639 "is_configured": true, 00:11:33.639 "data_offset": 2048, 00:11:33.639 "data_size": 63488 00:11:33.639 }, 00:11:33.639 { 00:11:33.639 "name": null, 00:11:33.639 "uuid": "2c5d827c-a20e-4ad8-b5c2-2ea145391ad4", 00:11:33.639 "is_configured": false, 00:11:33.639 "data_offset": 0, 00:11:33.639 "data_size": 63488 00:11:33.639 }, 00:11:33.639 { 00:11:33.639 "name": "BaseBdev3", 00:11:33.639 "uuid": "a0f2d327-f572-4c13-8054-9852c394411a", 00:11:33.639 "is_configured": true, 00:11:33.639 "data_offset": 2048, 00:11:33.639 "data_size": 63488 00:11:33.639 }, 00:11:33.639 { 00:11:33.639 "name": "BaseBdev4", 00:11:33.639 "uuid": "9fa36fa8-1027-45fe-97e9-5734edd38091", 00:11:33.639 "is_configured": true, 00:11:33.639 "data_offset": 2048, 00:11:33.639 "data_size": 63488 00:11:33.639 } 00:11:33.639 ] 00:11:33.639 }' 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:33.639 15:59:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.900 [2024-11-20 15:59:32.069376] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.900 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.160 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:34.160 "name": "Existed_Raid", 00:11:34.160 "uuid": "a92153ed-1a99-43d7-a82d-0e0dd960cdd3", 00:11:34.160 "strip_size_kb": 64, 00:11:34.160 "state": "configuring", 00:11:34.160 "raid_level": "concat", 00:11:34.160 "superblock": true, 00:11:34.160 "num_base_bdevs": 4, 00:11:34.160 "num_base_bdevs_discovered": 2, 00:11:34.160 "num_base_bdevs_operational": 4, 00:11:34.160 "base_bdevs_list": [ 00:11:34.160 { 00:11:34.160 "name": null, 00:11:34.160 "uuid": "730300c5-2cfd-470f-afbb-f4d249862836", 00:11:34.160 "is_configured": false, 00:11:34.160 "data_offset": 0, 00:11:34.160 "data_size": 63488 00:11:34.160 }, 00:11:34.160 { 00:11:34.160 "name": null, 00:11:34.160 "uuid": "2c5d827c-a20e-4ad8-b5c2-2ea145391ad4", 00:11:34.160 "is_configured": false, 00:11:34.160 "data_offset": 0, 00:11:34.160 "data_size": 63488 00:11:34.160 }, 00:11:34.160 { 00:11:34.160 "name": "BaseBdev3", 00:11:34.160 "uuid": "a0f2d327-f572-4c13-8054-9852c394411a", 00:11:34.160 "is_configured": true, 00:11:34.160 "data_offset": 2048, 00:11:34.160 "data_size": 63488 00:11:34.160 }, 00:11:34.160 { 00:11:34.160 "name": "BaseBdev4", 00:11:34.160 "uuid": "9fa36fa8-1027-45fe-97e9-5734edd38091", 00:11:34.160 "is_configured": true, 00:11:34.160 "data_offset": 2048, 00:11:34.160 "data_size": 63488 00:11:34.160 } 00:11:34.160 ] 00:11:34.160 }' 00:11:34.160 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:34.160 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.421 [2024-11-20 15:59:32.488682] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:34.421 "name": "Existed_Raid", 00:11:34.421 "uuid": "a92153ed-1a99-43d7-a82d-0e0dd960cdd3", 00:11:34.421 "strip_size_kb": 64, 00:11:34.421 "state": "configuring", 00:11:34.421 "raid_level": "concat", 00:11:34.421 "superblock": true, 00:11:34.421 "num_base_bdevs": 4, 00:11:34.421 "num_base_bdevs_discovered": 3, 00:11:34.421 "num_base_bdevs_operational": 4, 00:11:34.421 "base_bdevs_list": [ 00:11:34.421 { 00:11:34.421 "name": null, 00:11:34.421 "uuid": "730300c5-2cfd-470f-afbb-f4d249862836", 00:11:34.421 "is_configured": false, 00:11:34.421 "data_offset": 0, 00:11:34.421 "data_size": 63488 00:11:34.421 }, 00:11:34.421 { 00:11:34.421 "name": "BaseBdev2", 00:11:34.421 "uuid": "2c5d827c-a20e-4ad8-b5c2-2ea145391ad4", 00:11:34.421 "is_configured": true, 00:11:34.421 "data_offset": 2048, 00:11:34.421 "data_size": 63488 00:11:34.421 }, 00:11:34.421 { 00:11:34.421 "name": "BaseBdev3", 00:11:34.421 "uuid": "a0f2d327-f572-4c13-8054-9852c394411a", 00:11:34.421 "is_configured": true, 00:11:34.421 "data_offset": 2048, 00:11:34.421 "data_size": 63488 00:11:34.421 }, 00:11:34.421 { 00:11:34.421 "name": "BaseBdev4", 00:11:34.421 "uuid": "9fa36fa8-1027-45fe-97e9-5734edd38091", 00:11:34.421 "is_configured": true, 00:11:34.421 "data_offset": 2048, 00:11:34.421 "data_size": 63488 00:11:34.421 } 00:11:34.421 ] 00:11:34.421 }' 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:34.421 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.702 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:34.702 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 730300c5-2cfd-470f-afbb-f4d249862836 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.703 [2024-11-20 15:59:32.915011] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:11:34.703 NewBaseBdev 00:11:34.703 [2024-11-20 15:59:32.915335] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:11:34.703 [2024-11-20 15:59:32.915369] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:34.703 [2024-11-20 15:59:32.915621] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:11:34.703 [2024-11-20 15:59:32.915760] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:11:34.703 [2024-11-20 15:59:32.915772] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:11:34.703 [2024-11-20 15:59:32.915883] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.703 [ 00:11:34.703 { 00:11:34.703 "name": "NewBaseBdev", 00:11:34.703 "aliases": [ 00:11:34.703 "730300c5-2cfd-470f-afbb-f4d249862836" 00:11:34.703 ], 00:11:34.703 "product_name": "Malloc disk", 00:11:34.703 "block_size": 512, 00:11:34.703 "num_blocks": 65536, 00:11:34.703 "uuid": "730300c5-2cfd-470f-afbb-f4d249862836", 00:11:34.703 "assigned_rate_limits": { 00:11:34.703 "rw_ios_per_sec": 0, 00:11:34.703 "rw_mbytes_per_sec": 0, 00:11:34.703 "r_mbytes_per_sec": 0, 00:11:34.703 "w_mbytes_per_sec": 0 00:11:34.703 }, 00:11:34.703 "claimed": true, 00:11:34.703 "claim_type": "exclusive_write", 00:11:34.703 "zoned": false, 00:11:34.703 "supported_io_types": { 00:11:34.703 "read": true, 00:11:34.703 "write": true, 00:11:34.703 "unmap": true, 00:11:34.703 "flush": true, 00:11:34.703 "reset": true, 00:11:34.703 "nvme_admin": false, 00:11:34.703 "nvme_io": false, 00:11:34.703 "nvme_io_md": false, 00:11:34.703 "write_zeroes": true, 00:11:34.703 "zcopy": true, 00:11:34.703 "get_zone_info": false, 00:11:34.703 "zone_management": false, 00:11:34.703 "zone_append": false, 00:11:34.703 "compare": false, 00:11:34.703 "compare_and_write": false, 00:11:34.703 "abort": true, 00:11:34.703 "seek_hole": false, 00:11:34.703 "seek_data": false, 00:11:34.703 "copy": true, 00:11:34.703 "nvme_iov_md": false 00:11:34.703 }, 00:11:34.703 "memory_domains": [ 00:11:34.703 { 00:11:34.703 "dma_device_id": "system", 00:11:34.703 "dma_device_type": 1 00:11:34.703 }, 00:11:34.703 { 00:11:34.703 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:34.703 "dma_device_type": 2 00:11:34.703 } 00:11:34.703 ], 00:11:34.703 "driver_specific": {} 00:11:34.703 } 00:11:34.703 ] 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.703 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.965 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.965 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:34.965 "name": "Existed_Raid", 00:11:34.965 "uuid": "a92153ed-1a99-43d7-a82d-0e0dd960cdd3", 00:11:34.965 "strip_size_kb": 64, 00:11:34.965 "state": "online", 00:11:34.965 "raid_level": "concat", 00:11:34.965 "superblock": true, 00:11:34.965 "num_base_bdevs": 4, 00:11:34.965 "num_base_bdevs_discovered": 4, 00:11:34.965 "num_base_bdevs_operational": 4, 00:11:34.965 "base_bdevs_list": [ 00:11:34.965 { 00:11:34.965 "name": "NewBaseBdev", 00:11:34.965 "uuid": "730300c5-2cfd-470f-afbb-f4d249862836", 00:11:34.965 "is_configured": true, 00:11:34.965 "data_offset": 2048, 00:11:34.965 "data_size": 63488 00:11:34.965 }, 00:11:34.965 { 00:11:34.965 "name": "BaseBdev2", 00:11:34.965 "uuid": "2c5d827c-a20e-4ad8-b5c2-2ea145391ad4", 00:11:34.965 "is_configured": true, 00:11:34.965 "data_offset": 2048, 00:11:34.965 "data_size": 63488 00:11:34.965 }, 00:11:34.965 { 00:11:34.965 "name": "BaseBdev3", 00:11:34.965 "uuid": "a0f2d327-f572-4c13-8054-9852c394411a", 00:11:34.965 "is_configured": true, 00:11:34.965 "data_offset": 2048, 00:11:34.965 "data_size": 63488 00:11:34.965 }, 00:11:34.965 { 00:11:34.965 "name": "BaseBdev4", 00:11:34.965 "uuid": "9fa36fa8-1027-45fe-97e9-5734edd38091", 00:11:34.965 "is_configured": true, 00:11:34.965 "data_offset": 2048, 00:11:34.965 "data_size": 63488 00:11:34.965 } 00:11:34.965 ] 00:11:34.965 }' 00:11:34.965 15:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:34.965 15:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:35.226 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:11:35.226 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:35.226 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:35.226 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:35.226 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:11:35.226 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:35.226 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:35.226 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:35.226 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:35.226 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:35.226 [2024-11-20 15:59:33.259496] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:35.226 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:35.226 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:35.226 "name": "Existed_Raid", 00:11:35.226 "aliases": [ 00:11:35.226 "a92153ed-1a99-43d7-a82d-0e0dd960cdd3" 00:11:35.226 ], 00:11:35.226 "product_name": "Raid Volume", 00:11:35.226 "block_size": 512, 00:11:35.226 "num_blocks": 253952, 00:11:35.226 "uuid": "a92153ed-1a99-43d7-a82d-0e0dd960cdd3", 00:11:35.226 "assigned_rate_limits": { 00:11:35.226 "rw_ios_per_sec": 0, 00:11:35.226 "rw_mbytes_per_sec": 0, 00:11:35.226 "r_mbytes_per_sec": 0, 00:11:35.226 "w_mbytes_per_sec": 0 00:11:35.226 }, 00:11:35.226 "claimed": false, 00:11:35.226 "zoned": false, 00:11:35.226 "supported_io_types": { 00:11:35.226 "read": true, 00:11:35.226 "write": true, 00:11:35.226 "unmap": true, 00:11:35.226 "flush": true, 00:11:35.226 "reset": true, 00:11:35.226 "nvme_admin": false, 00:11:35.226 "nvme_io": false, 00:11:35.226 "nvme_io_md": false, 00:11:35.226 "write_zeroes": true, 00:11:35.226 "zcopy": false, 00:11:35.226 "get_zone_info": false, 00:11:35.226 "zone_management": false, 00:11:35.226 "zone_append": false, 00:11:35.226 "compare": false, 00:11:35.226 "compare_and_write": false, 00:11:35.226 "abort": false, 00:11:35.226 "seek_hole": false, 00:11:35.226 "seek_data": false, 00:11:35.226 "copy": false, 00:11:35.226 "nvme_iov_md": false 00:11:35.226 }, 00:11:35.226 "memory_domains": [ 00:11:35.226 { 00:11:35.226 "dma_device_id": "system", 00:11:35.226 "dma_device_type": 1 00:11:35.226 }, 00:11:35.226 { 00:11:35.226 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:35.226 "dma_device_type": 2 00:11:35.226 }, 00:11:35.226 { 00:11:35.226 "dma_device_id": "system", 00:11:35.226 "dma_device_type": 1 00:11:35.226 }, 00:11:35.226 { 00:11:35.226 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:35.226 "dma_device_type": 2 00:11:35.226 }, 00:11:35.226 { 00:11:35.226 "dma_device_id": "system", 00:11:35.226 "dma_device_type": 1 00:11:35.226 }, 00:11:35.226 { 00:11:35.226 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:35.226 "dma_device_type": 2 00:11:35.226 }, 00:11:35.226 { 00:11:35.226 "dma_device_id": "system", 00:11:35.226 "dma_device_type": 1 00:11:35.226 }, 00:11:35.226 { 00:11:35.226 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:35.226 "dma_device_type": 2 00:11:35.226 } 00:11:35.226 ], 00:11:35.226 "driver_specific": { 00:11:35.226 "raid": { 00:11:35.226 "uuid": "a92153ed-1a99-43d7-a82d-0e0dd960cdd3", 00:11:35.226 "strip_size_kb": 64, 00:11:35.226 "state": "online", 00:11:35.226 "raid_level": "concat", 00:11:35.226 "superblock": true, 00:11:35.226 "num_base_bdevs": 4, 00:11:35.226 "num_base_bdevs_discovered": 4, 00:11:35.226 "num_base_bdevs_operational": 4, 00:11:35.226 "base_bdevs_list": [ 00:11:35.226 { 00:11:35.226 "name": "NewBaseBdev", 00:11:35.226 "uuid": "730300c5-2cfd-470f-afbb-f4d249862836", 00:11:35.226 "is_configured": true, 00:11:35.226 "data_offset": 2048, 00:11:35.226 "data_size": 63488 00:11:35.226 }, 00:11:35.226 { 00:11:35.226 "name": "BaseBdev2", 00:11:35.226 "uuid": "2c5d827c-a20e-4ad8-b5c2-2ea145391ad4", 00:11:35.226 "is_configured": true, 00:11:35.226 "data_offset": 2048, 00:11:35.226 "data_size": 63488 00:11:35.226 }, 00:11:35.226 { 00:11:35.226 "name": "BaseBdev3", 00:11:35.226 "uuid": "a0f2d327-f572-4c13-8054-9852c394411a", 00:11:35.226 "is_configured": true, 00:11:35.226 "data_offset": 2048, 00:11:35.226 "data_size": 63488 00:11:35.226 }, 00:11:35.227 { 00:11:35.227 "name": "BaseBdev4", 00:11:35.227 "uuid": "9fa36fa8-1027-45fe-97e9-5734edd38091", 00:11:35.227 "is_configured": true, 00:11:35.227 "data_offset": 2048, 00:11:35.227 "data_size": 63488 00:11:35.227 } 00:11:35.227 ] 00:11:35.227 } 00:11:35.227 } 00:11:35.227 }' 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:11:35.227 BaseBdev2 00:11:35.227 BaseBdev3 00:11:35.227 BaseBdev4' 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:35.227 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:35.488 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:35.488 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:35.488 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:35.488 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:35.488 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:35.488 [2024-11-20 15:59:33.491185] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:35.488 [2024-11-20 15:59:33.491213] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:35.488 [2024-11-20 15:59:33.491283] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:35.488 [2024-11-20 15:59:33.491349] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:35.488 [2024-11-20 15:59:33.491358] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:11:35.488 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:35.488 15:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 69997 00:11:35.488 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 69997 ']' 00:11:35.488 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 69997 00:11:35.488 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:11:35.488 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:35.488 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69997 00:11:35.488 killing process with pid 69997 00:11:35.488 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:35.488 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:35.488 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69997' 00:11:35.488 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 69997 00:11:35.488 [2024-11-20 15:59:33.523650] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:35.488 15:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 69997 00:11:35.750 [2024-11-20 15:59:33.765576] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:36.321 ************************************ 00:11:36.321 END TEST raid_state_function_test_sb 00:11:36.321 ************************************ 00:11:36.321 15:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:11:36.321 00:11:36.321 real 0m8.486s 00:11:36.321 user 0m13.531s 00:11:36.321 sys 0m1.354s 00:11:36.321 15:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:36.321 15:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:36.321 15:59:34 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 4 00:11:36.321 15:59:34 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:11:36.321 15:59:34 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:36.321 15:59:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:36.321 ************************************ 00:11:36.321 START TEST raid_superblock_test 00:11:36.321 ************************************ 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 4 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=70634 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 70634 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 70634 ']' 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:11:36.322 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:36.322 15:59:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:36.584 [2024-11-20 15:59:34.602729] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:11:36.584 [2024-11-20 15:59:34.603002] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70634 ] 00:11:36.584 [2024-11-20 15:59:34.760896] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:36.844 [2024-11-20 15:59:34.863452] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:36.844 [2024-11-20 15:59:35.000615] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:36.844 [2024-11-20 15:59:35.000684] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.415 malloc1 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.415 [2024-11-20 15:59:35.521239] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:11:37.415 [2024-11-20 15:59:35.521300] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:37.415 [2024-11-20 15:59:35.521323] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:37.415 [2024-11-20 15:59:35.521333] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:37.415 [2024-11-20 15:59:35.523527] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:37.415 [2024-11-20 15:59:35.523563] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:11:37.415 pt1 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.415 malloc2 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.415 [2024-11-20 15:59:35.565325] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:37.415 [2024-11-20 15:59:35.565477] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:37.415 [2024-11-20 15:59:35.565505] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:37.415 [2024-11-20 15:59:35.565513] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:37.415 [2024-11-20 15:59:35.567619] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:37.415 [2024-11-20 15:59:35.567651] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:37.415 pt2 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:37.415 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.416 malloc3 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.416 [2024-11-20 15:59:35.620369] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:11:37.416 [2024-11-20 15:59:35.620420] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:37.416 [2024-11-20 15:59:35.620444] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:11:37.416 [2024-11-20 15:59:35.620453] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:37.416 [2024-11-20 15:59:35.622595] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:37.416 [2024-11-20 15:59:35.622630] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:11:37.416 pt3 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.416 malloc4 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.416 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.416 [2024-11-20 15:59:35.660472] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:37.416 [2024-11-20 15:59:35.660631] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:37.416 [2024-11-20 15:59:35.660656] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:11:37.416 [2024-11-20 15:59:35.660681] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:37.416 [2024-11-20 15:59:35.662782] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:37.416 [2024-11-20 15:59:35.662806] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:37.678 pt4 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.678 [2024-11-20 15:59:35.668512] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:11:37.678 [2024-11-20 15:59:35.670344] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:37.678 [2024-11-20 15:59:35.670417] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:37.678 [2024-11-20 15:59:35.670462] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:37.678 [2024-11-20 15:59:35.670634] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:11:37.678 [2024-11-20 15:59:35.670644] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:37.678 [2024-11-20 15:59:35.670917] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:11:37.678 [2024-11-20 15:59:35.671059] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:11:37.678 [2024-11-20 15:59:35.671070] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:11:37.678 [2024-11-20 15:59:35.671204] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:37.678 "name": "raid_bdev1", 00:11:37.678 "uuid": "d472b8eb-a901-4852-9ba0-948f9137943b", 00:11:37.678 "strip_size_kb": 64, 00:11:37.678 "state": "online", 00:11:37.678 "raid_level": "concat", 00:11:37.678 "superblock": true, 00:11:37.678 "num_base_bdevs": 4, 00:11:37.678 "num_base_bdevs_discovered": 4, 00:11:37.678 "num_base_bdevs_operational": 4, 00:11:37.678 "base_bdevs_list": [ 00:11:37.678 { 00:11:37.678 "name": "pt1", 00:11:37.678 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:37.678 "is_configured": true, 00:11:37.678 "data_offset": 2048, 00:11:37.678 "data_size": 63488 00:11:37.678 }, 00:11:37.678 { 00:11:37.678 "name": "pt2", 00:11:37.678 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:37.678 "is_configured": true, 00:11:37.678 "data_offset": 2048, 00:11:37.678 "data_size": 63488 00:11:37.678 }, 00:11:37.678 { 00:11:37.678 "name": "pt3", 00:11:37.678 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:37.678 "is_configured": true, 00:11:37.678 "data_offset": 2048, 00:11:37.678 "data_size": 63488 00:11:37.678 }, 00:11:37.678 { 00:11:37.678 "name": "pt4", 00:11:37.678 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:37.678 "is_configured": true, 00:11:37.678 "data_offset": 2048, 00:11:37.678 "data_size": 63488 00:11:37.678 } 00:11:37.678 ] 00:11:37.678 }' 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:37.678 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.938 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:11:37.938 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:11:37.938 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:37.938 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:37.938 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:37.938 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:37.938 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:37.938 15:59:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:37.938 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.938 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.938 [2024-11-20 15:59:35.984976] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:37.938 15:59:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.938 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:37.938 "name": "raid_bdev1", 00:11:37.938 "aliases": [ 00:11:37.938 "d472b8eb-a901-4852-9ba0-948f9137943b" 00:11:37.938 ], 00:11:37.938 "product_name": "Raid Volume", 00:11:37.938 "block_size": 512, 00:11:37.938 "num_blocks": 253952, 00:11:37.938 "uuid": "d472b8eb-a901-4852-9ba0-948f9137943b", 00:11:37.938 "assigned_rate_limits": { 00:11:37.938 "rw_ios_per_sec": 0, 00:11:37.938 "rw_mbytes_per_sec": 0, 00:11:37.938 "r_mbytes_per_sec": 0, 00:11:37.938 "w_mbytes_per_sec": 0 00:11:37.938 }, 00:11:37.938 "claimed": false, 00:11:37.938 "zoned": false, 00:11:37.938 "supported_io_types": { 00:11:37.938 "read": true, 00:11:37.938 "write": true, 00:11:37.938 "unmap": true, 00:11:37.938 "flush": true, 00:11:37.938 "reset": true, 00:11:37.938 "nvme_admin": false, 00:11:37.938 "nvme_io": false, 00:11:37.938 "nvme_io_md": false, 00:11:37.938 "write_zeroes": true, 00:11:37.938 "zcopy": false, 00:11:37.938 "get_zone_info": false, 00:11:37.938 "zone_management": false, 00:11:37.938 "zone_append": false, 00:11:37.938 "compare": false, 00:11:37.938 "compare_and_write": false, 00:11:37.938 "abort": false, 00:11:37.938 "seek_hole": false, 00:11:37.938 "seek_data": false, 00:11:37.938 "copy": false, 00:11:37.938 "nvme_iov_md": false 00:11:37.938 }, 00:11:37.938 "memory_domains": [ 00:11:37.938 { 00:11:37.938 "dma_device_id": "system", 00:11:37.938 "dma_device_type": 1 00:11:37.938 }, 00:11:37.938 { 00:11:37.938 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:37.938 "dma_device_type": 2 00:11:37.938 }, 00:11:37.938 { 00:11:37.938 "dma_device_id": "system", 00:11:37.938 "dma_device_type": 1 00:11:37.938 }, 00:11:37.938 { 00:11:37.938 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:37.938 "dma_device_type": 2 00:11:37.938 }, 00:11:37.938 { 00:11:37.938 "dma_device_id": "system", 00:11:37.938 "dma_device_type": 1 00:11:37.938 }, 00:11:37.938 { 00:11:37.938 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:37.938 "dma_device_type": 2 00:11:37.938 }, 00:11:37.938 { 00:11:37.938 "dma_device_id": "system", 00:11:37.938 "dma_device_type": 1 00:11:37.938 }, 00:11:37.938 { 00:11:37.938 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:37.938 "dma_device_type": 2 00:11:37.938 } 00:11:37.938 ], 00:11:37.938 "driver_specific": { 00:11:37.938 "raid": { 00:11:37.938 "uuid": "d472b8eb-a901-4852-9ba0-948f9137943b", 00:11:37.938 "strip_size_kb": 64, 00:11:37.938 "state": "online", 00:11:37.938 "raid_level": "concat", 00:11:37.938 "superblock": true, 00:11:37.938 "num_base_bdevs": 4, 00:11:37.938 "num_base_bdevs_discovered": 4, 00:11:37.938 "num_base_bdevs_operational": 4, 00:11:37.938 "base_bdevs_list": [ 00:11:37.938 { 00:11:37.938 "name": "pt1", 00:11:37.938 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:37.938 "is_configured": true, 00:11:37.938 "data_offset": 2048, 00:11:37.938 "data_size": 63488 00:11:37.938 }, 00:11:37.938 { 00:11:37.938 "name": "pt2", 00:11:37.938 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:37.938 "is_configured": true, 00:11:37.938 "data_offset": 2048, 00:11:37.938 "data_size": 63488 00:11:37.938 }, 00:11:37.938 { 00:11:37.938 "name": "pt3", 00:11:37.938 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:37.938 "is_configured": true, 00:11:37.938 "data_offset": 2048, 00:11:37.938 "data_size": 63488 00:11:37.938 }, 00:11:37.938 { 00:11:37.938 "name": "pt4", 00:11:37.938 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:37.938 "is_configured": true, 00:11:37.938 "data_offset": 2048, 00:11:37.938 "data_size": 63488 00:11:37.938 } 00:11:37.938 ] 00:11:37.938 } 00:11:37.938 } 00:11:37.938 }' 00:11:37.938 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:37.938 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:11:37.938 pt2 00:11:37.938 pt3 00:11:37.938 pt4' 00:11:37.938 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:37.938 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:37.938 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:37.938 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:37.938 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:11:37.938 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.938 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.938 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.939 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.197 [2024-11-20 15:59:36.216963] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=d472b8eb-a901-4852-9ba0-948f9137943b 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z d472b8eb-a901-4852-9ba0-948f9137943b ']' 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.197 [2024-11-20 15:59:36.240636] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:38.197 [2024-11-20 15:59:36.240661] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:38.197 [2024-11-20 15:59:36.240754] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:38.197 [2024-11-20 15:59:36.240827] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:38.197 [2024-11-20 15:59:36.240841] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.197 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.198 [2024-11-20 15:59:36.352689] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:11:38.198 [2024-11-20 15:59:36.354626] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:11:38.198 [2024-11-20 15:59:36.354687] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:11:38.198 [2024-11-20 15:59:36.354722] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:11:38.198 [2024-11-20 15:59:36.354767] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:11:38.198 [2024-11-20 15:59:36.354812] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:11:38.198 [2024-11-20 15:59:36.354831] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:11:38.198 [2024-11-20 15:59:36.354850] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:11:38.198 [2024-11-20 15:59:36.354862] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:38.198 [2024-11-20 15:59:36.354873] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:11:38.198 request: 00:11:38.198 { 00:11:38.198 "name": "raid_bdev1", 00:11:38.198 "raid_level": "concat", 00:11:38.198 "base_bdevs": [ 00:11:38.198 "malloc1", 00:11:38.198 "malloc2", 00:11:38.198 "malloc3", 00:11:38.198 "malloc4" 00:11:38.198 ], 00:11:38.198 "strip_size_kb": 64, 00:11:38.198 "superblock": false, 00:11:38.198 "method": "bdev_raid_create", 00:11:38.198 "req_id": 1 00:11:38.198 } 00:11:38.198 Got JSON-RPC error response 00:11:38.198 response: 00:11:38.198 { 00:11:38.198 "code": -17, 00:11:38.198 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:11:38.198 } 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.198 [2024-11-20 15:59:36.400687] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:11:38.198 [2024-11-20 15:59:36.400842] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:38.198 [2024-11-20 15:59:36.400880] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:38.198 [2024-11-20 15:59:36.400938] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:38.198 [2024-11-20 15:59:36.403126] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:38.198 [2024-11-20 15:59:36.403247] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:11:38.198 [2024-11-20 15:59:36.403370] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:11:38.198 [2024-11-20 15:59:36.403477] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:11:38.198 pt1 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:38.198 "name": "raid_bdev1", 00:11:38.198 "uuid": "d472b8eb-a901-4852-9ba0-948f9137943b", 00:11:38.198 "strip_size_kb": 64, 00:11:38.198 "state": "configuring", 00:11:38.198 "raid_level": "concat", 00:11:38.198 "superblock": true, 00:11:38.198 "num_base_bdevs": 4, 00:11:38.198 "num_base_bdevs_discovered": 1, 00:11:38.198 "num_base_bdevs_operational": 4, 00:11:38.198 "base_bdevs_list": [ 00:11:38.198 { 00:11:38.198 "name": "pt1", 00:11:38.198 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:38.198 "is_configured": true, 00:11:38.198 "data_offset": 2048, 00:11:38.198 "data_size": 63488 00:11:38.198 }, 00:11:38.198 { 00:11:38.198 "name": null, 00:11:38.198 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:38.198 "is_configured": false, 00:11:38.198 "data_offset": 2048, 00:11:38.198 "data_size": 63488 00:11:38.198 }, 00:11:38.198 { 00:11:38.198 "name": null, 00:11:38.198 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:38.198 "is_configured": false, 00:11:38.198 "data_offset": 2048, 00:11:38.198 "data_size": 63488 00:11:38.198 }, 00:11:38.198 { 00:11:38.198 "name": null, 00:11:38.198 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:38.198 "is_configured": false, 00:11:38.198 "data_offset": 2048, 00:11:38.198 "data_size": 63488 00:11:38.198 } 00:11:38.198 ] 00:11:38.198 }' 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:38.198 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.769 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:11:38.769 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:38.769 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.769 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.770 [2024-11-20 15:59:36.720768] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:38.770 [2024-11-20 15:59:36.720833] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:38.770 [2024-11-20 15:59:36.720850] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:11:38.770 [2024-11-20 15:59:36.720861] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:38.770 [2024-11-20 15:59:36.721257] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:38.770 [2024-11-20 15:59:36.721277] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:38.770 [2024-11-20 15:59:36.721345] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:11:38.770 [2024-11-20 15:59:36.721366] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:38.770 pt2 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.770 [2024-11-20 15:59:36.728772] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:38.770 "name": "raid_bdev1", 00:11:38.770 "uuid": "d472b8eb-a901-4852-9ba0-948f9137943b", 00:11:38.770 "strip_size_kb": 64, 00:11:38.770 "state": "configuring", 00:11:38.770 "raid_level": "concat", 00:11:38.770 "superblock": true, 00:11:38.770 "num_base_bdevs": 4, 00:11:38.770 "num_base_bdevs_discovered": 1, 00:11:38.770 "num_base_bdevs_operational": 4, 00:11:38.770 "base_bdevs_list": [ 00:11:38.770 { 00:11:38.770 "name": "pt1", 00:11:38.770 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:38.770 "is_configured": true, 00:11:38.770 "data_offset": 2048, 00:11:38.770 "data_size": 63488 00:11:38.770 }, 00:11:38.770 { 00:11:38.770 "name": null, 00:11:38.770 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:38.770 "is_configured": false, 00:11:38.770 "data_offset": 0, 00:11:38.770 "data_size": 63488 00:11:38.770 }, 00:11:38.770 { 00:11:38.770 "name": null, 00:11:38.770 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:38.770 "is_configured": false, 00:11:38.770 "data_offset": 2048, 00:11:38.770 "data_size": 63488 00:11:38.770 }, 00:11:38.770 { 00:11:38.770 "name": null, 00:11:38.770 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:38.770 "is_configured": false, 00:11:38.770 "data_offset": 2048, 00:11:38.770 "data_size": 63488 00:11:38.770 } 00:11:38.770 ] 00:11:38.770 }' 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:38.770 15:59:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.031 [2024-11-20 15:59:37.060848] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:39.031 [2024-11-20 15:59:37.061004] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:39.031 [2024-11-20 15:59:37.061043] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:11:39.031 [2024-11-20 15:59:37.061096] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:39.031 [2024-11-20 15:59:37.061511] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:39.031 [2024-11-20 15:59:37.061602] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:39.031 [2024-11-20 15:59:37.061736] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:11:39.031 [2024-11-20 15:59:37.061777] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:39.031 pt2 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.031 [2024-11-20 15:59:37.068835] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:11:39.031 [2024-11-20 15:59:37.068878] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:39.031 [2024-11-20 15:59:37.068895] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:11:39.031 [2024-11-20 15:59:37.068905] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:39.031 [2024-11-20 15:59:37.069271] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:39.031 [2024-11-20 15:59:37.069284] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:11:39.031 [2024-11-20 15:59:37.069349] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:11:39.031 [2024-11-20 15:59:37.069370] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:39.031 pt3 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.031 [2024-11-20 15:59:37.076815] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:39.031 [2024-11-20 15:59:37.076853] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:39.031 [2024-11-20 15:59:37.076868] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:11:39.031 [2024-11-20 15:59:37.076875] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:39.031 [2024-11-20 15:59:37.077213] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:39.031 [2024-11-20 15:59:37.077225] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:39.031 [2024-11-20 15:59:37.077279] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:11:39.031 [2024-11-20 15:59:37.077298] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:39.031 [2024-11-20 15:59:37.077422] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:11:39.031 [2024-11-20 15:59:37.077430] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:39.031 [2024-11-20 15:59:37.077682] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:11:39.031 [2024-11-20 15:59:37.077812] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:11:39.031 [2024-11-20 15:59:37.077823] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:11:39.031 [2024-11-20 15:59:37.077959] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:39.031 pt4 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:39.031 "name": "raid_bdev1", 00:11:39.031 "uuid": "d472b8eb-a901-4852-9ba0-948f9137943b", 00:11:39.031 "strip_size_kb": 64, 00:11:39.031 "state": "online", 00:11:39.031 "raid_level": "concat", 00:11:39.031 "superblock": true, 00:11:39.031 "num_base_bdevs": 4, 00:11:39.031 "num_base_bdevs_discovered": 4, 00:11:39.031 "num_base_bdevs_operational": 4, 00:11:39.031 "base_bdevs_list": [ 00:11:39.031 { 00:11:39.031 "name": "pt1", 00:11:39.031 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:39.031 "is_configured": true, 00:11:39.031 "data_offset": 2048, 00:11:39.031 "data_size": 63488 00:11:39.031 }, 00:11:39.031 { 00:11:39.031 "name": "pt2", 00:11:39.031 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:39.031 "is_configured": true, 00:11:39.031 "data_offset": 2048, 00:11:39.031 "data_size": 63488 00:11:39.031 }, 00:11:39.031 { 00:11:39.031 "name": "pt3", 00:11:39.031 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:39.031 "is_configured": true, 00:11:39.031 "data_offset": 2048, 00:11:39.031 "data_size": 63488 00:11:39.031 }, 00:11:39.031 { 00:11:39.031 "name": "pt4", 00:11:39.031 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:39.031 "is_configured": true, 00:11:39.031 "data_offset": 2048, 00:11:39.031 "data_size": 63488 00:11:39.031 } 00:11:39.031 ] 00:11:39.031 }' 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:39.031 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.290 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:11:39.290 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:11:39.290 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:39.290 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:39.290 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:39.290 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:39.290 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:39.290 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:39.290 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.290 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.290 [2024-11-20 15:59:37.413270] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:39.290 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.290 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:39.290 "name": "raid_bdev1", 00:11:39.290 "aliases": [ 00:11:39.290 "d472b8eb-a901-4852-9ba0-948f9137943b" 00:11:39.290 ], 00:11:39.290 "product_name": "Raid Volume", 00:11:39.290 "block_size": 512, 00:11:39.290 "num_blocks": 253952, 00:11:39.290 "uuid": "d472b8eb-a901-4852-9ba0-948f9137943b", 00:11:39.290 "assigned_rate_limits": { 00:11:39.290 "rw_ios_per_sec": 0, 00:11:39.291 "rw_mbytes_per_sec": 0, 00:11:39.291 "r_mbytes_per_sec": 0, 00:11:39.291 "w_mbytes_per_sec": 0 00:11:39.291 }, 00:11:39.291 "claimed": false, 00:11:39.291 "zoned": false, 00:11:39.291 "supported_io_types": { 00:11:39.291 "read": true, 00:11:39.291 "write": true, 00:11:39.291 "unmap": true, 00:11:39.291 "flush": true, 00:11:39.291 "reset": true, 00:11:39.291 "nvme_admin": false, 00:11:39.291 "nvme_io": false, 00:11:39.291 "nvme_io_md": false, 00:11:39.291 "write_zeroes": true, 00:11:39.291 "zcopy": false, 00:11:39.291 "get_zone_info": false, 00:11:39.291 "zone_management": false, 00:11:39.291 "zone_append": false, 00:11:39.291 "compare": false, 00:11:39.291 "compare_and_write": false, 00:11:39.291 "abort": false, 00:11:39.291 "seek_hole": false, 00:11:39.291 "seek_data": false, 00:11:39.291 "copy": false, 00:11:39.291 "nvme_iov_md": false 00:11:39.291 }, 00:11:39.291 "memory_domains": [ 00:11:39.291 { 00:11:39.291 "dma_device_id": "system", 00:11:39.291 "dma_device_type": 1 00:11:39.291 }, 00:11:39.291 { 00:11:39.291 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:39.291 "dma_device_type": 2 00:11:39.291 }, 00:11:39.291 { 00:11:39.291 "dma_device_id": "system", 00:11:39.291 "dma_device_type": 1 00:11:39.291 }, 00:11:39.291 { 00:11:39.291 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:39.291 "dma_device_type": 2 00:11:39.291 }, 00:11:39.291 { 00:11:39.291 "dma_device_id": "system", 00:11:39.291 "dma_device_type": 1 00:11:39.291 }, 00:11:39.291 { 00:11:39.291 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:39.291 "dma_device_type": 2 00:11:39.291 }, 00:11:39.291 { 00:11:39.291 "dma_device_id": "system", 00:11:39.291 "dma_device_type": 1 00:11:39.291 }, 00:11:39.291 { 00:11:39.291 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:39.291 "dma_device_type": 2 00:11:39.291 } 00:11:39.291 ], 00:11:39.291 "driver_specific": { 00:11:39.291 "raid": { 00:11:39.291 "uuid": "d472b8eb-a901-4852-9ba0-948f9137943b", 00:11:39.291 "strip_size_kb": 64, 00:11:39.291 "state": "online", 00:11:39.291 "raid_level": "concat", 00:11:39.291 "superblock": true, 00:11:39.291 "num_base_bdevs": 4, 00:11:39.291 "num_base_bdevs_discovered": 4, 00:11:39.291 "num_base_bdevs_operational": 4, 00:11:39.291 "base_bdevs_list": [ 00:11:39.291 { 00:11:39.291 "name": "pt1", 00:11:39.291 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:39.291 "is_configured": true, 00:11:39.291 "data_offset": 2048, 00:11:39.291 "data_size": 63488 00:11:39.291 }, 00:11:39.291 { 00:11:39.291 "name": "pt2", 00:11:39.291 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:39.291 "is_configured": true, 00:11:39.291 "data_offset": 2048, 00:11:39.291 "data_size": 63488 00:11:39.291 }, 00:11:39.291 { 00:11:39.291 "name": "pt3", 00:11:39.291 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:39.291 "is_configured": true, 00:11:39.291 "data_offset": 2048, 00:11:39.291 "data_size": 63488 00:11:39.291 }, 00:11:39.291 { 00:11:39.291 "name": "pt4", 00:11:39.291 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:39.291 "is_configured": true, 00:11:39.291 "data_offset": 2048, 00:11:39.291 "data_size": 63488 00:11:39.291 } 00:11:39.291 ] 00:11:39.291 } 00:11:39.291 } 00:11:39.291 }' 00:11:39.291 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:39.291 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:11:39.291 pt2 00:11:39.291 pt3 00:11:39.291 pt4' 00:11:39.291 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:39.291 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:39.291 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:39.291 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:11:39.291 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:39.291 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.291 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.552 [2024-11-20 15:59:37.677299] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' d472b8eb-a901-4852-9ba0-948f9137943b '!=' d472b8eb-a901-4852-9ba0-948f9137943b ']' 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 70634 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 70634 ']' 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 70634 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70634 00:11:39.552 killing process with pid 70634 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70634' 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 70634 00:11:39.552 [2024-11-20 15:59:37.734676] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:39.552 [2024-11-20 15:59:37.734748] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:39.552 15:59:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 70634 00:11:39.552 [2024-11-20 15:59:37.734821] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:39.552 [2024-11-20 15:59:37.734831] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:11:39.814 [2024-11-20 15:59:37.978264] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:40.751 15:59:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:11:40.751 00:11:40.751 real 0m4.152s 00:11:40.751 user 0m6.003s 00:11:40.751 sys 0m0.624s 00:11:40.751 ************************************ 00:11:40.751 END TEST raid_superblock_test 00:11:40.751 ************************************ 00:11:40.751 15:59:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:40.751 15:59:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:40.751 15:59:38 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 4 read 00:11:40.751 15:59:38 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:40.751 15:59:38 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:40.751 15:59:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:40.751 ************************************ 00:11:40.751 START TEST raid_read_error_test 00:11:40.751 ************************************ 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 4 read 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.wWo1lGVO59 00:11:40.751 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=70882 00:11:40.752 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 70882 00:11:40.752 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:40.752 15:59:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 70882 ']' 00:11:40.752 15:59:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:40.752 15:59:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:11:40.752 15:59:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:40.752 15:59:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:40.752 15:59:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:40.752 15:59:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:40.752 [2024-11-20 15:59:38.807421] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:11:40.752 [2024-11-20 15:59:38.807541] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70882 ] 00:11:40.752 [2024-11-20 15:59:38.965699] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:41.008 [2024-11-20 15:59:39.065930] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:41.008 [2024-11-20 15:59:39.202602] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:41.008 [2024-11-20 15:59:39.202634] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:41.579 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:41.579 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:11:41.579 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:41.579 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:41.579 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.579 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.579 BaseBdev1_malloc 00:11:41.579 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.579 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:11:41.579 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.579 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.579 true 00:11:41.579 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.579 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:41.579 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.579 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.579 [2024-11-20 15:59:39.708336] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:41.579 [2024-11-20 15:59:39.708387] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:41.579 [2024-11-20 15:59:39.708407] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:11:41.579 [2024-11-20 15:59:39.708419] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:41.579 [2024-11-20 15:59:39.710568] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:41.579 [2024-11-20 15:59:39.710607] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:41.579 BaseBdev1 00:11:41.579 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.579 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:41.579 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.580 BaseBdev2_malloc 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.580 true 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.580 [2024-11-20 15:59:39.760553] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:41.580 [2024-11-20 15:59:39.760736] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:41.580 [2024-11-20 15:59:39.760761] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:41.580 [2024-11-20 15:59:39.760772] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:41.580 [2024-11-20 15:59:39.762907] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:41.580 [2024-11-20 15:59:39.762940] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:41.580 BaseBdev2 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.580 BaseBdev3_malloc 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.580 true 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.580 [2024-11-20 15:59:39.817622] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:11:41.580 [2024-11-20 15:59:39.817690] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:41.580 [2024-11-20 15:59:39.817709] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:11:41.580 [2024-11-20 15:59:39.817719] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:41.580 [2024-11-20 15:59:39.819872] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:41.580 [2024-11-20 15:59:39.819906] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:41.580 BaseBdev3 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.580 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.842 BaseBdev4_malloc 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.842 true 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.842 [2024-11-20 15:59:39.861744] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:11:41.842 [2024-11-20 15:59:39.861793] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:41.842 [2024-11-20 15:59:39.861813] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:41.842 [2024-11-20 15:59:39.861824] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:41.842 [2024-11-20 15:59:39.863957] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:41.842 [2024-11-20 15:59:39.863993] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:41.842 BaseBdev4 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.842 [2024-11-20 15:59:39.869793] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:41.842 [2024-11-20 15:59:39.871651] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:41.842 [2024-11-20 15:59:39.871740] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:41.842 [2024-11-20 15:59:39.871807] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:41.842 [2024-11-20 15:59:39.872024] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:11:41.842 [2024-11-20 15:59:39.872036] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:41.842 [2024-11-20 15:59:39.872281] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:11:41.842 [2024-11-20 15:59:39.872422] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:11:41.842 [2024-11-20 15:59:39.872432] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:11:41.842 [2024-11-20 15:59:39.872570] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:41.842 "name": "raid_bdev1", 00:11:41.842 "uuid": "b9cd491d-c5d2-4bfe-bffe-2370cfe09ae1", 00:11:41.842 "strip_size_kb": 64, 00:11:41.842 "state": "online", 00:11:41.842 "raid_level": "concat", 00:11:41.842 "superblock": true, 00:11:41.842 "num_base_bdevs": 4, 00:11:41.842 "num_base_bdevs_discovered": 4, 00:11:41.842 "num_base_bdevs_operational": 4, 00:11:41.842 "base_bdevs_list": [ 00:11:41.842 { 00:11:41.842 "name": "BaseBdev1", 00:11:41.842 "uuid": "e737ffdc-c1c6-51e9-b28e-32721f7062c2", 00:11:41.842 "is_configured": true, 00:11:41.842 "data_offset": 2048, 00:11:41.842 "data_size": 63488 00:11:41.842 }, 00:11:41.842 { 00:11:41.842 "name": "BaseBdev2", 00:11:41.842 "uuid": "12a079da-bf46-5f42-a830-396cea0e954d", 00:11:41.842 "is_configured": true, 00:11:41.842 "data_offset": 2048, 00:11:41.842 "data_size": 63488 00:11:41.842 }, 00:11:41.842 { 00:11:41.842 "name": "BaseBdev3", 00:11:41.842 "uuid": "86828b0c-ad5c-552c-8955-df059c414e14", 00:11:41.842 "is_configured": true, 00:11:41.842 "data_offset": 2048, 00:11:41.842 "data_size": 63488 00:11:41.842 }, 00:11:41.842 { 00:11:41.842 "name": "BaseBdev4", 00:11:41.842 "uuid": "88646143-453d-5758-8ec1-5f109c660a73", 00:11:41.842 "is_configured": true, 00:11:41.842 "data_offset": 2048, 00:11:41.842 "data_size": 63488 00:11:41.842 } 00:11:41.842 ] 00:11:41.842 }' 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:41.842 15:59:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:42.102 15:59:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:11:42.102 15:59:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:42.102 [2024-11-20 15:59:40.282818] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006a40 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:43.092 "name": "raid_bdev1", 00:11:43.092 "uuid": "b9cd491d-c5d2-4bfe-bffe-2370cfe09ae1", 00:11:43.092 "strip_size_kb": 64, 00:11:43.092 "state": "online", 00:11:43.092 "raid_level": "concat", 00:11:43.092 "superblock": true, 00:11:43.092 "num_base_bdevs": 4, 00:11:43.092 "num_base_bdevs_discovered": 4, 00:11:43.092 "num_base_bdevs_operational": 4, 00:11:43.092 "base_bdevs_list": [ 00:11:43.092 { 00:11:43.092 "name": "BaseBdev1", 00:11:43.092 "uuid": "e737ffdc-c1c6-51e9-b28e-32721f7062c2", 00:11:43.092 "is_configured": true, 00:11:43.092 "data_offset": 2048, 00:11:43.092 "data_size": 63488 00:11:43.092 }, 00:11:43.092 { 00:11:43.092 "name": "BaseBdev2", 00:11:43.092 "uuid": "12a079da-bf46-5f42-a830-396cea0e954d", 00:11:43.092 "is_configured": true, 00:11:43.092 "data_offset": 2048, 00:11:43.092 "data_size": 63488 00:11:43.092 }, 00:11:43.092 { 00:11:43.092 "name": "BaseBdev3", 00:11:43.092 "uuid": "86828b0c-ad5c-552c-8955-df059c414e14", 00:11:43.092 "is_configured": true, 00:11:43.092 "data_offset": 2048, 00:11:43.092 "data_size": 63488 00:11:43.092 }, 00:11:43.092 { 00:11:43.092 "name": "BaseBdev4", 00:11:43.092 "uuid": "88646143-453d-5758-8ec1-5f109c660a73", 00:11:43.092 "is_configured": true, 00:11:43.092 "data_offset": 2048, 00:11:43.092 "data_size": 63488 00:11:43.092 } 00:11:43.092 ] 00:11:43.092 }' 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:43.092 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:43.352 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:43.352 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:43.352 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:43.352 [2024-11-20 15:59:41.545312] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:43.352 [2024-11-20 15:59:41.545483] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:43.352 [2024-11-20 15:59:41.548577] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:43.352 [2024-11-20 15:59:41.548758] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:43.352 [2024-11-20 15:59:41.548811] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:43.352 [2024-11-20 15:59:41.548824] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:11:43.352 { 00:11:43.352 "results": [ 00:11:43.352 { 00:11:43.352 "job": "raid_bdev1", 00:11:43.352 "core_mask": "0x1", 00:11:43.352 "workload": "randrw", 00:11:43.352 "percentage": 50, 00:11:43.352 "status": "finished", 00:11:43.352 "queue_depth": 1, 00:11:43.352 "io_size": 131072, 00:11:43.352 "runtime": 1.26071, 00:11:43.352 "iops": 14512.457266143681, 00:11:43.352 "mibps": 1814.0571582679602, 00:11:43.352 "io_failed": 1, 00:11:43.352 "io_timeout": 0, 00:11:43.352 "avg_latency_us": 93.97567016030371, 00:11:43.352 "min_latency_us": 33.673846153846156, 00:11:43.352 "max_latency_us": 1688.8123076923077 00:11:43.352 } 00:11:43.352 ], 00:11:43.352 "core_count": 1 00:11:43.352 } 00:11:43.352 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:43.352 15:59:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 70882 00:11:43.352 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 70882 ']' 00:11:43.352 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 70882 00:11:43.352 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:11:43.352 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:43.352 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70882 00:11:43.352 killing process with pid 70882 00:11:43.352 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:43.352 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:43.352 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70882' 00:11:43.352 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 70882 00:11:43.352 [2024-11-20 15:59:41.577969] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:43.352 15:59:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 70882 00:11:43.612 [2024-11-20 15:59:41.779187] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:44.551 15:59:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.wWo1lGVO59 00:11:44.551 15:59:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:11:44.551 15:59:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:11:44.551 15:59:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.79 00:11:44.551 15:59:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:11:44.551 15:59:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:44.551 15:59:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:11:44.551 15:59:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.79 != \0\.\0\0 ]] 00:11:44.551 00:11:44.551 real 0m3.882s 00:11:44.551 user 0m4.578s 00:11:44.551 sys 0m0.404s 00:11:44.551 15:59:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:44.551 15:59:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:44.551 ************************************ 00:11:44.551 END TEST raid_read_error_test 00:11:44.551 ************************************ 00:11:44.551 15:59:42 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 4 write 00:11:44.551 15:59:42 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:44.551 15:59:42 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:44.551 15:59:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:44.551 ************************************ 00:11:44.551 START TEST raid_write_error_test 00:11:44.551 ************************************ 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 4 write 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.SsIC1408d5 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=71022 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 71022 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 71022 ']' 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:44.551 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:44.551 15:59:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:44.551 [2024-11-20 15:59:42.756614] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:11:44.551 [2024-11-20 15:59:42.756757] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71022 ] 00:11:44.812 [2024-11-20 15:59:42.912153] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:44.812 [2024-11-20 15:59:43.018345] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:45.073 [2024-11-20 15:59:43.159388] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:45.073 [2024-11-20 15:59:43.159437] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.642 BaseBdev1_malloc 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.642 true 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.642 [2024-11-20 15:59:43.676911] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:45.642 [2024-11-20 15:59:43.676964] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:45.642 [2024-11-20 15:59:43.676986] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:11:45.642 [2024-11-20 15:59:43.676997] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:45.642 [2024-11-20 15:59:43.679175] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:45.642 [2024-11-20 15:59:43.679213] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:45.642 BaseBdev1 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.642 BaseBdev2_malloc 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.642 true 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.642 [2024-11-20 15:59:43.721126] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:45.642 [2024-11-20 15:59:43.721178] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:45.642 [2024-11-20 15:59:43.721195] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:45.642 [2024-11-20 15:59:43.721205] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:45.642 [2024-11-20 15:59:43.723362] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:45.642 [2024-11-20 15:59:43.723401] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:45.642 BaseBdev2 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.642 BaseBdev3_malloc 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.642 true 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.642 [2024-11-20 15:59:43.779187] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:11:45.642 [2024-11-20 15:59:43.779243] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:45.642 [2024-11-20 15:59:43.779262] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:11:45.642 [2024-11-20 15:59:43.779274] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:45.642 [2024-11-20 15:59:43.781402] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:45.642 [2024-11-20 15:59:43.781438] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:45.642 BaseBdev3 00:11:45.642 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.643 BaseBdev4_malloc 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.643 true 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.643 [2024-11-20 15:59:43.824282] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:11:45.643 [2024-11-20 15:59:43.824335] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:45.643 [2024-11-20 15:59:43.824355] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:45.643 [2024-11-20 15:59:43.824366] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:45.643 [2024-11-20 15:59:43.826506] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:45.643 [2024-11-20 15:59:43.826545] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:45.643 BaseBdev4 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.643 [2024-11-20 15:59:43.832342] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:45.643 [2024-11-20 15:59:43.834188] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:45.643 [2024-11-20 15:59:43.834268] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:45.643 [2024-11-20 15:59:43.834335] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:45.643 [2024-11-20 15:59:43.834560] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:11:45.643 [2024-11-20 15:59:43.834579] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:45.643 [2024-11-20 15:59:43.834837] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:11:45.643 [2024-11-20 15:59:43.834989] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:11:45.643 [2024-11-20 15:59:43.835006] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:11:45.643 [2024-11-20 15:59:43.835148] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:45.643 "name": "raid_bdev1", 00:11:45.643 "uuid": "e0e7ea21-483a-4e85-901f-590d7316768b", 00:11:45.643 "strip_size_kb": 64, 00:11:45.643 "state": "online", 00:11:45.643 "raid_level": "concat", 00:11:45.643 "superblock": true, 00:11:45.643 "num_base_bdevs": 4, 00:11:45.643 "num_base_bdevs_discovered": 4, 00:11:45.643 "num_base_bdevs_operational": 4, 00:11:45.643 "base_bdevs_list": [ 00:11:45.643 { 00:11:45.643 "name": "BaseBdev1", 00:11:45.643 "uuid": "1760b061-bbc3-53c7-aaa0-e3210913ccb2", 00:11:45.643 "is_configured": true, 00:11:45.643 "data_offset": 2048, 00:11:45.643 "data_size": 63488 00:11:45.643 }, 00:11:45.643 { 00:11:45.643 "name": "BaseBdev2", 00:11:45.643 "uuid": "f8bd334b-6132-5574-b5c9-00764086993f", 00:11:45.643 "is_configured": true, 00:11:45.643 "data_offset": 2048, 00:11:45.643 "data_size": 63488 00:11:45.643 }, 00:11:45.643 { 00:11:45.643 "name": "BaseBdev3", 00:11:45.643 "uuid": "fd4a5be7-1074-57b9-823f-6650e271f8bc", 00:11:45.643 "is_configured": true, 00:11:45.643 "data_offset": 2048, 00:11:45.643 "data_size": 63488 00:11:45.643 }, 00:11:45.643 { 00:11:45.643 "name": "BaseBdev4", 00:11:45.643 "uuid": "fd6c0d21-8da8-5978-b4d2-50980ed112f1", 00:11:45.643 "is_configured": true, 00:11:45.643 "data_offset": 2048, 00:11:45.643 "data_size": 63488 00:11:45.643 } 00:11:45.643 ] 00:11:45.643 }' 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:45.643 15:59:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:46.214 15:59:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:11:46.214 15:59:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:46.214 [2024-11-20 15:59:44.245369] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006a40 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:47.156 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:47.157 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:47.157 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:47.157 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:47.157 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:47.157 "name": "raid_bdev1", 00:11:47.157 "uuid": "e0e7ea21-483a-4e85-901f-590d7316768b", 00:11:47.157 "strip_size_kb": 64, 00:11:47.157 "state": "online", 00:11:47.157 "raid_level": "concat", 00:11:47.157 "superblock": true, 00:11:47.157 "num_base_bdevs": 4, 00:11:47.157 "num_base_bdevs_discovered": 4, 00:11:47.157 "num_base_bdevs_operational": 4, 00:11:47.157 "base_bdevs_list": [ 00:11:47.157 { 00:11:47.157 "name": "BaseBdev1", 00:11:47.157 "uuid": "1760b061-bbc3-53c7-aaa0-e3210913ccb2", 00:11:47.157 "is_configured": true, 00:11:47.157 "data_offset": 2048, 00:11:47.157 "data_size": 63488 00:11:47.157 }, 00:11:47.157 { 00:11:47.157 "name": "BaseBdev2", 00:11:47.157 "uuid": "f8bd334b-6132-5574-b5c9-00764086993f", 00:11:47.157 "is_configured": true, 00:11:47.157 "data_offset": 2048, 00:11:47.157 "data_size": 63488 00:11:47.157 }, 00:11:47.157 { 00:11:47.157 "name": "BaseBdev3", 00:11:47.157 "uuid": "fd4a5be7-1074-57b9-823f-6650e271f8bc", 00:11:47.157 "is_configured": true, 00:11:47.157 "data_offset": 2048, 00:11:47.157 "data_size": 63488 00:11:47.157 }, 00:11:47.157 { 00:11:47.157 "name": "BaseBdev4", 00:11:47.157 "uuid": "fd6c0d21-8da8-5978-b4d2-50980ed112f1", 00:11:47.157 "is_configured": true, 00:11:47.157 "data_offset": 2048, 00:11:47.157 "data_size": 63488 00:11:47.157 } 00:11:47.157 ] 00:11:47.157 }' 00:11:47.157 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:47.157 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:47.416 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:47.416 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:47.416 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:47.416 [2024-11-20 15:59:45.475074] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:47.416 [2024-11-20 15:59:45.475107] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:47.416 [2024-11-20 15:59:45.478182] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:47.416 [2024-11-20 15:59:45.478250] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:47.416 [2024-11-20 15:59:45.478295] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:47.416 [2024-11-20 15:59:45.478308] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:11:47.416 { 00:11:47.416 "results": [ 00:11:47.416 { 00:11:47.416 "job": "raid_bdev1", 00:11:47.416 "core_mask": "0x1", 00:11:47.416 "workload": "randrw", 00:11:47.416 "percentage": 50, 00:11:47.416 "status": "finished", 00:11:47.416 "queue_depth": 1, 00:11:47.416 "io_size": 131072, 00:11:47.416 "runtime": 1.227836, 00:11:47.416 "iops": 14362.66732690685, 00:11:47.416 "mibps": 1795.3334158633563, 00:11:47.416 "io_failed": 1, 00:11:47.416 "io_timeout": 0, 00:11:47.416 "avg_latency_us": 95.00082977127204, 00:11:47.416 "min_latency_us": 33.673846153846156, 00:11:47.416 "max_latency_us": 1688.8123076923077 00:11:47.416 } 00:11:47.416 ], 00:11:47.416 "core_count": 1 00:11:47.416 } 00:11:47.416 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:47.416 15:59:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 71022 00:11:47.416 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 71022 ']' 00:11:47.416 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 71022 00:11:47.416 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:11:47.416 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:47.416 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71022 00:11:47.416 killing process with pid 71022 00:11:47.416 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:47.416 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:47.416 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71022' 00:11:47.416 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 71022 00:11:47.416 15:59:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 71022 00:11:47.416 [2024-11-20 15:59:45.504002] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:47.676 [2024-11-20 15:59:45.707823] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:48.244 15:59:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.SsIC1408d5 00:11:48.244 15:59:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:11:48.244 15:59:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:11:48.244 15:59:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.81 00:11:48.244 15:59:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:11:48.244 15:59:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:48.244 15:59:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:11:48.244 15:59:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.81 != \0\.\0\0 ]] 00:11:48.244 00:11:48.244 real 0m3.783s 00:11:48.244 user 0m4.488s 00:11:48.244 sys 0m0.404s 00:11:48.244 15:59:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:48.244 15:59:46 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:48.244 ************************************ 00:11:48.244 END TEST raid_write_error_test 00:11:48.244 ************************************ 00:11:48.504 15:59:46 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:11:48.504 15:59:46 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 4 false 00:11:48.504 15:59:46 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:48.504 15:59:46 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:48.504 15:59:46 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:48.504 ************************************ 00:11:48.504 START TEST raid_state_function_test 00:11:48.504 ************************************ 00:11:48.504 15:59:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 4 false 00:11:48.504 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:11:48.504 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:11:48.504 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:11:48.504 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:11:48.505 Process raid pid: 71155 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=71155 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 71155' 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 71155 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 71155 ']' 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:48.505 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:48.505 15:59:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:11:48.505 [2024-11-20 15:59:46.591410] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:11:48.505 [2024-11-20 15:59:46.591535] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:48.505 [2024-11-20 15:59:46.751472] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:48.766 [2024-11-20 15:59:46.854820] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:48.766 [2024-11-20 15:59:46.993810] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:48.766 [2024-11-20 15:59:46.993854] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.397 [2024-11-20 15:59:47.446455] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:49.397 [2024-11-20 15:59:47.446508] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:49.397 [2024-11-20 15:59:47.446519] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:49.397 [2024-11-20 15:59:47.446530] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:49.397 [2024-11-20 15:59:47.446541] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:49.397 [2024-11-20 15:59:47.446551] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:49.397 [2024-11-20 15:59:47.446557] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:49.397 [2024-11-20 15:59:47.446567] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:49.397 "name": "Existed_Raid", 00:11:49.397 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:49.397 "strip_size_kb": 0, 00:11:49.397 "state": "configuring", 00:11:49.397 "raid_level": "raid1", 00:11:49.397 "superblock": false, 00:11:49.397 "num_base_bdevs": 4, 00:11:49.397 "num_base_bdevs_discovered": 0, 00:11:49.397 "num_base_bdevs_operational": 4, 00:11:49.397 "base_bdevs_list": [ 00:11:49.397 { 00:11:49.397 "name": "BaseBdev1", 00:11:49.397 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:49.397 "is_configured": false, 00:11:49.397 "data_offset": 0, 00:11:49.397 "data_size": 0 00:11:49.397 }, 00:11:49.397 { 00:11:49.397 "name": "BaseBdev2", 00:11:49.397 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:49.397 "is_configured": false, 00:11:49.397 "data_offset": 0, 00:11:49.397 "data_size": 0 00:11:49.397 }, 00:11:49.397 { 00:11:49.397 "name": "BaseBdev3", 00:11:49.397 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:49.397 "is_configured": false, 00:11:49.397 "data_offset": 0, 00:11:49.397 "data_size": 0 00:11:49.397 }, 00:11:49.397 { 00:11:49.397 "name": "BaseBdev4", 00:11:49.397 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:49.397 "is_configured": false, 00:11:49.397 "data_offset": 0, 00:11:49.397 "data_size": 0 00:11:49.397 } 00:11:49.397 ] 00:11:49.397 }' 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:49.397 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.659 [2024-11-20 15:59:47.770471] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:49.659 [2024-11-20 15:59:47.770510] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.659 [2024-11-20 15:59:47.778473] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:49.659 [2024-11-20 15:59:47.778512] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:49.659 [2024-11-20 15:59:47.778520] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:49.659 [2024-11-20 15:59:47.778529] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:49.659 [2024-11-20 15:59:47.778536] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:49.659 [2024-11-20 15:59:47.778545] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:49.659 [2024-11-20 15:59:47.778551] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:49.659 [2024-11-20 15:59:47.778560] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.659 [2024-11-20 15:59:47.811184] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:49.659 BaseBdev1 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.659 [ 00:11:49.659 { 00:11:49.659 "name": "BaseBdev1", 00:11:49.659 "aliases": [ 00:11:49.659 "2f73fe77-3d11-4f78-8d6f-ebf6a5f947d3" 00:11:49.659 ], 00:11:49.659 "product_name": "Malloc disk", 00:11:49.659 "block_size": 512, 00:11:49.659 "num_blocks": 65536, 00:11:49.659 "uuid": "2f73fe77-3d11-4f78-8d6f-ebf6a5f947d3", 00:11:49.659 "assigned_rate_limits": { 00:11:49.659 "rw_ios_per_sec": 0, 00:11:49.659 "rw_mbytes_per_sec": 0, 00:11:49.659 "r_mbytes_per_sec": 0, 00:11:49.659 "w_mbytes_per_sec": 0 00:11:49.659 }, 00:11:49.659 "claimed": true, 00:11:49.659 "claim_type": "exclusive_write", 00:11:49.659 "zoned": false, 00:11:49.659 "supported_io_types": { 00:11:49.659 "read": true, 00:11:49.659 "write": true, 00:11:49.659 "unmap": true, 00:11:49.659 "flush": true, 00:11:49.659 "reset": true, 00:11:49.659 "nvme_admin": false, 00:11:49.659 "nvme_io": false, 00:11:49.659 "nvme_io_md": false, 00:11:49.659 "write_zeroes": true, 00:11:49.659 "zcopy": true, 00:11:49.659 "get_zone_info": false, 00:11:49.659 "zone_management": false, 00:11:49.659 "zone_append": false, 00:11:49.659 "compare": false, 00:11:49.659 "compare_and_write": false, 00:11:49.659 "abort": true, 00:11:49.659 "seek_hole": false, 00:11:49.659 "seek_data": false, 00:11:49.659 "copy": true, 00:11:49.659 "nvme_iov_md": false 00:11:49.659 }, 00:11:49.659 "memory_domains": [ 00:11:49.659 { 00:11:49.659 "dma_device_id": "system", 00:11:49.659 "dma_device_type": 1 00:11:49.659 }, 00:11:49.659 { 00:11:49.659 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:49.659 "dma_device_type": 2 00:11:49.659 } 00:11:49.659 ], 00:11:49.659 "driver_specific": {} 00:11:49.659 } 00:11:49.659 ] 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:49.659 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:49.660 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.660 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.660 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:49.660 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.660 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:49.660 "name": "Existed_Raid", 00:11:49.660 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:49.660 "strip_size_kb": 0, 00:11:49.660 "state": "configuring", 00:11:49.660 "raid_level": "raid1", 00:11:49.660 "superblock": false, 00:11:49.660 "num_base_bdevs": 4, 00:11:49.660 "num_base_bdevs_discovered": 1, 00:11:49.660 "num_base_bdevs_operational": 4, 00:11:49.660 "base_bdevs_list": [ 00:11:49.660 { 00:11:49.660 "name": "BaseBdev1", 00:11:49.660 "uuid": "2f73fe77-3d11-4f78-8d6f-ebf6a5f947d3", 00:11:49.660 "is_configured": true, 00:11:49.660 "data_offset": 0, 00:11:49.660 "data_size": 65536 00:11:49.660 }, 00:11:49.660 { 00:11:49.660 "name": "BaseBdev2", 00:11:49.660 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:49.660 "is_configured": false, 00:11:49.660 "data_offset": 0, 00:11:49.660 "data_size": 0 00:11:49.660 }, 00:11:49.660 { 00:11:49.660 "name": "BaseBdev3", 00:11:49.660 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:49.660 "is_configured": false, 00:11:49.660 "data_offset": 0, 00:11:49.660 "data_size": 0 00:11:49.660 }, 00:11:49.660 { 00:11:49.660 "name": "BaseBdev4", 00:11:49.660 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:49.660 "is_configured": false, 00:11:49.660 "data_offset": 0, 00:11:49.660 "data_size": 0 00:11:49.660 } 00:11:49.660 ] 00:11:49.660 }' 00:11:49.660 15:59:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:49.660 15:59:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.920 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:49.920 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.920 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.920 [2024-11-20 15:59:48.147283] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:49.920 [2024-11-20 15:59:48.147330] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:11:49.920 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.920 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.921 [2024-11-20 15:59:48.155332] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:49.921 [2024-11-20 15:59:48.157208] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:49.921 [2024-11-20 15:59:48.157251] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:49.921 [2024-11-20 15:59:48.157260] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:49.921 [2024-11-20 15:59:48.157271] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:49.921 [2024-11-20 15:59:48.157277] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:49.921 [2024-11-20 15:59:48.157285] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:49.921 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.182 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.182 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:50.182 "name": "Existed_Raid", 00:11:50.182 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.182 "strip_size_kb": 0, 00:11:50.182 "state": "configuring", 00:11:50.182 "raid_level": "raid1", 00:11:50.182 "superblock": false, 00:11:50.182 "num_base_bdevs": 4, 00:11:50.182 "num_base_bdevs_discovered": 1, 00:11:50.182 "num_base_bdevs_operational": 4, 00:11:50.182 "base_bdevs_list": [ 00:11:50.182 { 00:11:50.182 "name": "BaseBdev1", 00:11:50.182 "uuid": "2f73fe77-3d11-4f78-8d6f-ebf6a5f947d3", 00:11:50.182 "is_configured": true, 00:11:50.182 "data_offset": 0, 00:11:50.182 "data_size": 65536 00:11:50.182 }, 00:11:50.182 { 00:11:50.182 "name": "BaseBdev2", 00:11:50.182 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.182 "is_configured": false, 00:11:50.182 "data_offset": 0, 00:11:50.182 "data_size": 0 00:11:50.182 }, 00:11:50.182 { 00:11:50.182 "name": "BaseBdev3", 00:11:50.182 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.182 "is_configured": false, 00:11:50.182 "data_offset": 0, 00:11:50.182 "data_size": 0 00:11:50.182 }, 00:11:50.182 { 00:11:50.182 "name": "BaseBdev4", 00:11:50.182 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.182 "is_configured": false, 00:11:50.182 "data_offset": 0, 00:11:50.182 "data_size": 0 00:11:50.182 } 00:11:50.182 ] 00:11:50.182 }' 00:11:50.182 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:50.182 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.444 [2024-11-20 15:59:48.498089] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:50.444 BaseBdev2 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.444 [ 00:11:50.444 { 00:11:50.444 "name": "BaseBdev2", 00:11:50.444 "aliases": [ 00:11:50.444 "15e76072-febf-4e41-bba9-26c654535c3d" 00:11:50.444 ], 00:11:50.444 "product_name": "Malloc disk", 00:11:50.444 "block_size": 512, 00:11:50.444 "num_blocks": 65536, 00:11:50.444 "uuid": "15e76072-febf-4e41-bba9-26c654535c3d", 00:11:50.444 "assigned_rate_limits": { 00:11:50.444 "rw_ios_per_sec": 0, 00:11:50.444 "rw_mbytes_per_sec": 0, 00:11:50.444 "r_mbytes_per_sec": 0, 00:11:50.444 "w_mbytes_per_sec": 0 00:11:50.444 }, 00:11:50.444 "claimed": true, 00:11:50.444 "claim_type": "exclusive_write", 00:11:50.444 "zoned": false, 00:11:50.444 "supported_io_types": { 00:11:50.444 "read": true, 00:11:50.444 "write": true, 00:11:50.444 "unmap": true, 00:11:50.444 "flush": true, 00:11:50.444 "reset": true, 00:11:50.444 "nvme_admin": false, 00:11:50.444 "nvme_io": false, 00:11:50.444 "nvme_io_md": false, 00:11:50.444 "write_zeroes": true, 00:11:50.444 "zcopy": true, 00:11:50.444 "get_zone_info": false, 00:11:50.444 "zone_management": false, 00:11:50.444 "zone_append": false, 00:11:50.444 "compare": false, 00:11:50.444 "compare_and_write": false, 00:11:50.444 "abort": true, 00:11:50.444 "seek_hole": false, 00:11:50.444 "seek_data": false, 00:11:50.444 "copy": true, 00:11:50.444 "nvme_iov_md": false 00:11:50.444 }, 00:11:50.444 "memory_domains": [ 00:11:50.444 { 00:11:50.444 "dma_device_id": "system", 00:11:50.444 "dma_device_type": 1 00:11:50.444 }, 00:11:50.444 { 00:11:50.444 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:50.444 "dma_device_type": 2 00:11:50.444 } 00:11:50.444 ], 00:11:50.444 "driver_specific": {} 00:11:50.444 } 00:11:50.444 ] 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:50.444 "name": "Existed_Raid", 00:11:50.444 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.444 "strip_size_kb": 0, 00:11:50.444 "state": "configuring", 00:11:50.444 "raid_level": "raid1", 00:11:50.444 "superblock": false, 00:11:50.444 "num_base_bdevs": 4, 00:11:50.444 "num_base_bdevs_discovered": 2, 00:11:50.444 "num_base_bdevs_operational": 4, 00:11:50.444 "base_bdevs_list": [ 00:11:50.444 { 00:11:50.444 "name": "BaseBdev1", 00:11:50.444 "uuid": "2f73fe77-3d11-4f78-8d6f-ebf6a5f947d3", 00:11:50.444 "is_configured": true, 00:11:50.444 "data_offset": 0, 00:11:50.444 "data_size": 65536 00:11:50.444 }, 00:11:50.444 { 00:11:50.444 "name": "BaseBdev2", 00:11:50.444 "uuid": "15e76072-febf-4e41-bba9-26c654535c3d", 00:11:50.444 "is_configured": true, 00:11:50.444 "data_offset": 0, 00:11:50.444 "data_size": 65536 00:11:50.444 }, 00:11:50.444 { 00:11:50.444 "name": "BaseBdev3", 00:11:50.444 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.444 "is_configured": false, 00:11:50.444 "data_offset": 0, 00:11:50.444 "data_size": 0 00:11:50.444 }, 00:11:50.444 { 00:11:50.444 "name": "BaseBdev4", 00:11:50.444 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.444 "is_configured": false, 00:11:50.444 "data_offset": 0, 00:11:50.444 "data_size": 0 00:11:50.444 } 00:11:50.444 ] 00:11:50.444 }' 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:50.444 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.703 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:50.703 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.703 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.704 [2024-11-20 15:59:48.916504] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:50.704 BaseBdev3 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.704 [ 00:11:50.704 { 00:11:50.704 "name": "BaseBdev3", 00:11:50.704 "aliases": [ 00:11:50.704 "f4e2da26-a7c0-4671-a45f-06e907890d85" 00:11:50.704 ], 00:11:50.704 "product_name": "Malloc disk", 00:11:50.704 "block_size": 512, 00:11:50.704 "num_blocks": 65536, 00:11:50.704 "uuid": "f4e2da26-a7c0-4671-a45f-06e907890d85", 00:11:50.704 "assigned_rate_limits": { 00:11:50.704 "rw_ios_per_sec": 0, 00:11:50.704 "rw_mbytes_per_sec": 0, 00:11:50.704 "r_mbytes_per_sec": 0, 00:11:50.704 "w_mbytes_per_sec": 0 00:11:50.704 }, 00:11:50.704 "claimed": true, 00:11:50.704 "claim_type": "exclusive_write", 00:11:50.704 "zoned": false, 00:11:50.704 "supported_io_types": { 00:11:50.704 "read": true, 00:11:50.704 "write": true, 00:11:50.704 "unmap": true, 00:11:50.704 "flush": true, 00:11:50.704 "reset": true, 00:11:50.704 "nvme_admin": false, 00:11:50.704 "nvme_io": false, 00:11:50.704 "nvme_io_md": false, 00:11:50.704 "write_zeroes": true, 00:11:50.704 "zcopy": true, 00:11:50.704 "get_zone_info": false, 00:11:50.704 "zone_management": false, 00:11:50.704 "zone_append": false, 00:11:50.704 "compare": false, 00:11:50.704 "compare_and_write": false, 00:11:50.704 "abort": true, 00:11:50.704 "seek_hole": false, 00:11:50.704 "seek_data": false, 00:11:50.704 "copy": true, 00:11:50.704 "nvme_iov_md": false 00:11:50.704 }, 00:11:50.704 "memory_domains": [ 00:11:50.704 { 00:11:50.704 "dma_device_id": "system", 00:11:50.704 "dma_device_type": 1 00:11:50.704 }, 00:11:50.704 { 00:11:50.704 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:50.704 "dma_device_type": 2 00:11:50.704 } 00:11:50.704 ], 00:11:50.704 "driver_specific": {} 00:11:50.704 } 00:11:50.704 ] 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.704 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.963 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.963 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:50.963 "name": "Existed_Raid", 00:11:50.963 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.963 "strip_size_kb": 0, 00:11:50.963 "state": "configuring", 00:11:50.963 "raid_level": "raid1", 00:11:50.963 "superblock": false, 00:11:50.963 "num_base_bdevs": 4, 00:11:50.963 "num_base_bdevs_discovered": 3, 00:11:50.963 "num_base_bdevs_operational": 4, 00:11:50.963 "base_bdevs_list": [ 00:11:50.963 { 00:11:50.963 "name": "BaseBdev1", 00:11:50.963 "uuid": "2f73fe77-3d11-4f78-8d6f-ebf6a5f947d3", 00:11:50.963 "is_configured": true, 00:11:50.963 "data_offset": 0, 00:11:50.963 "data_size": 65536 00:11:50.963 }, 00:11:50.963 { 00:11:50.963 "name": "BaseBdev2", 00:11:50.963 "uuid": "15e76072-febf-4e41-bba9-26c654535c3d", 00:11:50.963 "is_configured": true, 00:11:50.963 "data_offset": 0, 00:11:50.963 "data_size": 65536 00:11:50.963 }, 00:11:50.963 { 00:11:50.963 "name": "BaseBdev3", 00:11:50.963 "uuid": "f4e2da26-a7c0-4671-a45f-06e907890d85", 00:11:50.963 "is_configured": true, 00:11:50.963 "data_offset": 0, 00:11:50.963 "data_size": 65536 00:11:50.963 }, 00:11:50.963 { 00:11:50.963 "name": "BaseBdev4", 00:11:50.963 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.963 "is_configured": false, 00:11:50.963 "data_offset": 0, 00:11:50.963 "data_size": 0 00:11:50.963 } 00:11:50.963 ] 00:11:50.963 }' 00:11:50.963 15:59:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:50.963 15:59:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.224 [2024-11-20 15:59:49.287494] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:51.224 [2024-11-20 15:59:49.287547] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:11:51.224 [2024-11-20 15:59:49.287556] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:51.224 [2024-11-20 15:59:49.287830] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:11:51.224 [2024-11-20 15:59:49.287985] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:11:51.224 [2024-11-20 15:59:49.288002] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:11:51.224 [2024-11-20 15:59:49.288232] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:51.224 BaseBdev4 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.224 [ 00:11:51.224 { 00:11:51.224 "name": "BaseBdev4", 00:11:51.224 "aliases": [ 00:11:51.224 "761f8ec5-9e3b-47ab-b3f5-47d4357b205e" 00:11:51.224 ], 00:11:51.224 "product_name": "Malloc disk", 00:11:51.224 "block_size": 512, 00:11:51.224 "num_blocks": 65536, 00:11:51.224 "uuid": "761f8ec5-9e3b-47ab-b3f5-47d4357b205e", 00:11:51.224 "assigned_rate_limits": { 00:11:51.224 "rw_ios_per_sec": 0, 00:11:51.224 "rw_mbytes_per_sec": 0, 00:11:51.224 "r_mbytes_per_sec": 0, 00:11:51.224 "w_mbytes_per_sec": 0 00:11:51.224 }, 00:11:51.224 "claimed": true, 00:11:51.224 "claim_type": "exclusive_write", 00:11:51.224 "zoned": false, 00:11:51.224 "supported_io_types": { 00:11:51.224 "read": true, 00:11:51.224 "write": true, 00:11:51.224 "unmap": true, 00:11:51.224 "flush": true, 00:11:51.224 "reset": true, 00:11:51.224 "nvme_admin": false, 00:11:51.224 "nvme_io": false, 00:11:51.224 "nvme_io_md": false, 00:11:51.224 "write_zeroes": true, 00:11:51.224 "zcopy": true, 00:11:51.224 "get_zone_info": false, 00:11:51.224 "zone_management": false, 00:11:51.224 "zone_append": false, 00:11:51.224 "compare": false, 00:11:51.224 "compare_and_write": false, 00:11:51.224 "abort": true, 00:11:51.224 "seek_hole": false, 00:11:51.224 "seek_data": false, 00:11:51.224 "copy": true, 00:11:51.224 "nvme_iov_md": false 00:11:51.224 }, 00:11:51.224 "memory_domains": [ 00:11:51.224 { 00:11:51.224 "dma_device_id": "system", 00:11:51.224 "dma_device_type": 1 00:11:51.224 }, 00:11:51.224 { 00:11:51.224 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:51.224 "dma_device_type": 2 00:11:51.224 } 00:11:51.224 ], 00:11:51.224 "driver_specific": {} 00:11:51.224 } 00:11:51.224 ] 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:51.224 "name": "Existed_Raid", 00:11:51.224 "uuid": "57b35218-c7aa-4135-915c-e7005e5f1c3c", 00:11:51.224 "strip_size_kb": 0, 00:11:51.224 "state": "online", 00:11:51.224 "raid_level": "raid1", 00:11:51.224 "superblock": false, 00:11:51.224 "num_base_bdevs": 4, 00:11:51.224 "num_base_bdevs_discovered": 4, 00:11:51.224 "num_base_bdevs_operational": 4, 00:11:51.224 "base_bdevs_list": [ 00:11:51.224 { 00:11:51.224 "name": "BaseBdev1", 00:11:51.224 "uuid": "2f73fe77-3d11-4f78-8d6f-ebf6a5f947d3", 00:11:51.224 "is_configured": true, 00:11:51.224 "data_offset": 0, 00:11:51.224 "data_size": 65536 00:11:51.224 }, 00:11:51.224 { 00:11:51.224 "name": "BaseBdev2", 00:11:51.224 "uuid": "15e76072-febf-4e41-bba9-26c654535c3d", 00:11:51.224 "is_configured": true, 00:11:51.224 "data_offset": 0, 00:11:51.224 "data_size": 65536 00:11:51.224 }, 00:11:51.224 { 00:11:51.224 "name": "BaseBdev3", 00:11:51.224 "uuid": "f4e2da26-a7c0-4671-a45f-06e907890d85", 00:11:51.224 "is_configured": true, 00:11:51.224 "data_offset": 0, 00:11:51.224 "data_size": 65536 00:11:51.224 }, 00:11:51.224 { 00:11:51.224 "name": "BaseBdev4", 00:11:51.224 "uuid": "761f8ec5-9e3b-47ab-b3f5-47d4357b205e", 00:11:51.224 "is_configured": true, 00:11:51.224 "data_offset": 0, 00:11:51.224 "data_size": 65536 00:11:51.224 } 00:11:51.224 ] 00:11:51.224 }' 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:51.224 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.484 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:11:51.484 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:51.484 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:51.484 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:51.484 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:51.484 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:51.484 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:51.484 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:51.484 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.484 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.484 [2024-11-20 15:59:49.631986] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:51.484 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.484 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:51.484 "name": "Existed_Raid", 00:11:51.484 "aliases": [ 00:11:51.484 "57b35218-c7aa-4135-915c-e7005e5f1c3c" 00:11:51.484 ], 00:11:51.484 "product_name": "Raid Volume", 00:11:51.484 "block_size": 512, 00:11:51.484 "num_blocks": 65536, 00:11:51.484 "uuid": "57b35218-c7aa-4135-915c-e7005e5f1c3c", 00:11:51.484 "assigned_rate_limits": { 00:11:51.484 "rw_ios_per_sec": 0, 00:11:51.484 "rw_mbytes_per_sec": 0, 00:11:51.484 "r_mbytes_per_sec": 0, 00:11:51.484 "w_mbytes_per_sec": 0 00:11:51.484 }, 00:11:51.484 "claimed": false, 00:11:51.484 "zoned": false, 00:11:51.484 "supported_io_types": { 00:11:51.485 "read": true, 00:11:51.485 "write": true, 00:11:51.485 "unmap": false, 00:11:51.485 "flush": false, 00:11:51.485 "reset": true, 00:11:51.485 "nvme_admin": false, 00:11:51.485 "nvme_io": false, 00:11:51.485 "nvme_io_md": false, 00:11:51.485 "write_zeroes": true, 00:11:51.485 "zcopy": false, 00:11:51.485 "get_zone_info": false, 00:11:51.485 "zone_management": false, 00:11:51.485 "zone_append": false, 00:11:51.485 "compare": false, 00:11:51.485 "compare_and_write": false, 00:11:51.485 "abort": false, 00:11:51.485 "seek_hole": false, 00:11:51.485 "seek_data": false, 00:11:51.485 "copy": false, 00:11:51.485 "nvme_iov_md": false 00:11:51.485 }, 00:11:51.485 "memory_domains": [ 00:11:51.485 { 00:11:51.485 "dma_device_id": "system", 00:11:51.485 "dma_device_type": 1 00:11:51.485 }, 00:11:51.485 { 00:11:51.485 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:51.485 "dma_device_type": 2 00:11:51.485 }, 00:11:51.485 { 00:11:51.485 "dma_device_id": "system", 00:11:51.485 "dma_device_type": 1 00:11:51.485 }, 00:11:51.485 { 00:11:51.485 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:51.485 "dma_device_type": 2 00:11:51.485 }, 00:11:51.485 { 00:11:51.485 "dma_device_id": "system", 00:11:51.485 "dma_device_type": 1 00:11:51.485 }, 00:11:51.485 { 00:11:51.485 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:51.485 "dma_device_type": 2 00:11:51.485 }, 00:11:51.485 { 00:11:51.485 "dma_device_id": "system", 00:11:51.485 "dma_device_type": 1 00:11:51.485 }, 00:11:51.485 { 00:11:51.485 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:51.485 "dma_device_type": 2 00:11:51.485 } 00:11:51.485 ], 00:11:51.485 "driver_specific": { 00:11:51.485 "raid": { 00:11:51.485 "uuid": "57b35218-c7aa-4135-915c-e7005e5f1c3c", 00:11:51.485 "strip_size_kb": 0, 00:11:51.485 "state": "online", 00:11:51.485 "raid_level": "raid1", 00:11:51.485 "superblock": false, 00:11:51.485 "num_base_bdevs": 4, 00:11:51.485 "num_base_bdevs_discovered": 4, 00:11:51.485 "num_base_bdevs_operational": 4, 00:11:51.485 "base_bdevs_list": [ 00:11:51.485 { 00:11:51.485 "name": "BaseBdev1", 00:11:51.485 "uuid": "2f73fe77-3d11-4f78-8d6f-ebf6a5f947d3", 00:11:51.485 "is_configured": true, 00:11:51.485 "data_offset": 0, 00:11:51.485 "data_size": 65536 00:11:51.485 }, 00:11:51.485 { 00:11:51.485 "name": "BaseBdev2", 00:11:51.485 "uuid": "15e76072-febf-4e41-bba9-26c654535c3d", 00:11:51.485 "is_configured": true, 00:11:51.485 "data_offset": 0, 00:11:51.485 "data_size": 65536 00:11:51.485 }, 00:11:51.485 { 00:11:51.485 "name": "BaseBdev3", 00:11:51.485 "uuid": "f4e2da26-a7c0-4671-a45f-06e907890d85", 00:11:51.485 "is_configured": true, 00:11:51.485 "data_offset": 0, 00:11:51.485 "data_size": 65536 00:11:51.485 }, 00:11:51.485 { 00:11:51.485 "name": "BaseBdev4", 00:11:51.485 "uuid": "761f8ec5-9e3b-47ab-b3f5-47d4357b205e", 00:11:51.485 "is_configured": true, 00:11:51.485 "data_offset": 0, 00:11:51.485 "data_size": 65536 00:11:51.485 } 00:11:51.485 ] 00:11:51.485 } 00:11:51.485 } 00:11:51.485 }' 00:11:51.485 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:51.485 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:11:51.485 BaseBdev2 00:11:51.485 BaseBdev3 00:11:51.485 BaseBdev4' 00:11:51.485 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:51.485 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:51.485 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:51.485 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:11:51.485 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:51.485 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.485 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.485 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.745 [2024-11-20 15:59:49.863748] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.745 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:51.745 "name": "Existed_Raid", 00:11:51.745 "uuid": "57b35218-c7aa-4135-915c-e7005e5f1c3c", 00:11:51.745 "strip_size_kb": 0, 00:11:51.745 "state": "online", 00:11:51.746 "raid_level": "raid1", 00:11:51.746 "superblock": false, 00:11:51.746 "num_base_bdevs": 4, 00:11:51.746 "num_base_bdevs_discovered": 3, 00:11:51.746 "num_base_bdevs_operational": 3, 00:11:51.746 "base_bdevs_list": [ 00:11:51.746 { 00:11:51.746 "name": null, 00:11:51.746 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:51.746 "is_configured": false, 00:11:51.746 "data_offset": 0, 00:11:51.746 "data_size": 65536 00:11:51.746 }, 00:11:51.746 { 00:11:51.746 "name": "BaseBdev2", 00:11:51.746 "uuid": "15e76072-febf-4e41-bba9-26c654535c3d", 00:11:51.746 "is_configured": true, 00:11:51.746 "data_offset": 0, 00:11:51.746 "data_size": 65536 00:11:51.746 }, 00:11:51.746 { 00:11:51.746 "name": "BaseBdev3", 00:11:51.746 "uuid": "f4e2da26-a7c0-4671-a45f-06e907890d85", 00:11:51.746 "is_configured": true, 00:11:51.746 "data_offset": 0, 00:11:51.746 "data_size": 65536 00:11:51.746 }, 00:11:51.746 { 00:11:51.746 "name": "BaseBdev4", 00:11:51.746 "uuid": "761f8ec5-9e3b-47ab-b3f5-47d4357b205e", 00:11:51.746 "is_configured": true, 00:11:51.746 "data_offset": 0, 00:11:51.746 "data_size": 65536 00:11:51.746 } 00:11:51.746 ] 00:11:51.746 }' 00:11:51.746 15:59:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:51.746 15:59:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.005 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:11:52.005 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:52.005 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.005 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.005 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.005 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.265 [2024-11-20 15:59:50.270788] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.265 [2024-11-20 15:59:50.369520] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:52.265 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.266 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:52.266 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:52.266 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.266 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:52.266 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.266 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.266 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.266 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:52.266 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:52.266 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:11:52.266 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.266 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.266 [2024-11-20 15:59:50.464889] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:11:52.266 [2024-11-20 15:59:50.465085] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:52.528 [2024-11-20 15:59:50.523513] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:52.528 [2024-11-20 15:59:50.523565] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:52.528 [2024-11-20 15:59:50.523576] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.528 BaseBdev2 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.528 [ 00:11:52.528 { 00:11:52.528 "name": "BaseBdev2", 00:11:52.528 "aliases": [ 00:11:52.528 "7bedd1d4-b5db-4f3b-8e8c-7097104aae97" 00:11:52.528 ], 00:11:52.528 "product_name": "Malloc disk", 00:11:52.528 "block_size": 512, 00:11:52.528 "num_blocks": 65536, 00:11:52.528 "uuid": "7bedd1d4-b5db-4f3b-8e8c-7097104aae97", 00:11:52.528 "assigned_rate_limits": { 00:11:52.528 "rw_ios_per_sec": 0, 00:11:52.528 "rw_mbytes_per_sec": 0, 00:11:52.528 "r_mbytes_per_sec": 0, 00:11:52.528 "w_mbytes_per_sec": 0 00:11:52.528 }, 00:11:52.528 "claimed": false, 00:11:52.528 "zoned": false, 00:11:52.528 "supported_io_types": { 00:11:52.528 "read": true, 00:11:52.528 "write": true, 00:11:52.528 "unmap": true, 00:11:52.528 "flush": true, 00:11:52.528 "reset": true, 00:11:52.528 "nvme_admin": false, 00:11:52.528 "nvme_io": false, 00:11:52.528 "nvme_io_md": false, 00:11:52.528 "write_zeroes": true, 00:11:52.528 "zcopy": true, 00:11:52.528 "get_zone_info": false, 00:11:52.528 "zone_management": false, 00:11:52.528 "zone_append": false, 00:11:52.528 "compare": false, 00:11:52.528 "compare_and_write": false, 00:11:52.528 "abort": true, 00:11:52.528 "seek_hole": false, 00:11:52.528 "seek_data": false, 00:11:52.528 "copy": true, 00:11:52.528 "nvme_iov_md": false 00:11:52.528 }, 00:11:52.528 "memory_domains": [ 00:11:52.528 { 00:11:52.528 "dma_device_id": "system", 00:11:52.528 "dma_device_type": 1 00:11:52.528 }, 00:11:52.528 { 00:11:52.528 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:52.528 "dma_device_type": 2 00:11:52.528 } 00:11:52.528 ], 00:11:52.528 "driver_specific": {} 00:11:52.528 } 00:11:52.528 ] 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.528 BaseBdev3 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.528 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.528 [ 00:11:52.528 { 00:11:52.528 "name": "BaseBdev3", 00:11:52.528 "aliases": [ 00:11:52.528 "b92061de-c461-41e2-b990-a9624a0c9d9a" 00:11:52.528 ], 00:11:52.528 "product_name": "Malloc disk", 00:11:52.528 "block_size": 512, 00:11:52.528 "num_blocks": 65536, 00:11:52.528 "uuid": "b92061de-c461-41e2-b990-a9624a0c9d9a", 00:11:52.528 "assigned_rate_limits": { 00:11:52.528 "rw_ios_per_sec": 0, 00:11:52.528 "rw_mbytes_per_sec": 0, 00:11:52.528 "r_mbytes_per_sec": 0, 00:11:52.528 "w_mbytes_per_sec": 0 00:11:52.528 }, 00:11:52.528 "claimed": false, 00:11:52.528 "zoned": false, 00:11:52.528 "supported_io_types": { 00:11:52.528 "read": true, 00:11:52.528 "write": true, 00:11:52.528 "unmap": true, 00:11:52.528 "flush": true, 00:11:52.529 "reset": true, 00:11:52.529 "nvme_admin": false, 00:11:52.529 "nvme_io": false, 00:11:52.529 "nvme_io_md": false, 00:11:52.529 "write_zeroes": true, 00:11:52.529 "zcopy": true, 00:11:52.529 "get_zone_info": false, 00:11:52.529 "zone_management": false, 00:11:52.529 "zone_append": false, 00:11:52.529 "compare": false, 00:11:52.529 "compare_and_write": false, 00:11:52.529 "abort": true, 00:11:52.529 "seek_hole": false, 00:11:52.529 "seek_data": false, 00:11:52.529 "copy": true, 00:11:52.529 "nvme_iov_md": false 00:11:52.529 }, 00:11:52.529 "memory_domains": [ 00:11:52.529 { 00:11:52.529 "dma_device_id": "system", 00:11:52.529 "dma_device_type": 1 00:11:52.529 }, 00:11:52.529 { 00:11:52.529 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:52.529 "dma_device_type": 2 00:11:52.529 } 00:11:52.529 ], 00:11:52.529 "driver_specific": {} 00:11:52.529 } 00:11:52.529 ] 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.529 BaseBdev4 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.529 [ 00:11:52.529 { 00:11:52.529 "name": "BaseBdev4", 00:11:52.529 "aliases": [ 00:11:52.529 "0455bb9f-a2fa-40ca-8a67-5cd915a4f050" 00:11:52.529 ], 00:11:52.529 "product_name": "Malloc disk", 00:11:52.529 "block_size": 512, 00:11:52.529 "num_blocks": 65536, 00:11:52.529 "uuid": "0455bb9f-a2fa-40ca-8a67-5cd915a4f050", 00:11:52.529 "assigned_rate_limits": { 00:11:52.529 "rw_ios_per_sec": 0, 00:11:52.529 "rw_mbytes_per_sec": 0, 00:11:52.529 "r_mbytes_per_sec": 0, 00:11:52.529 "w_mbytes_per_sec": 0 00:11:52.529 }, 00:11:52.529 "claimed": false, 00:11:52.529 "zoned": false, 00:11:52.529 "supported_io_types": { 00:11:52.529 "read": true, 00:11:52.529 "write": true, 00:11:52.529 "unmap": true, 00:11:52.529 "flush": true, 00:11:52.529 "reset": true, 00:11:52.529 "nvme_admin": false, 00:11:52.529 "nvme_io": false, 00:11:52.529 "nvme_io_md": false, 00:11:52.529 "write_zeroes": true, 00:11:52.529 "zcopy": true, 00:11:52.529 "get_zone_info": false, 00:11:52.529 "zone_management": false, 00:11:52.529 "zone_append": false, 00:11:52.529 "compare": false, 00:11:52.529 "compare_and_write": false, 00:11:52.529 "abort": true, 00:11:52.529 "seek_hole": false, 00:11:52.529 "seek_data": false, 00:11:52.529 "copy": true, 00:11:52.529 "nvme_iov_md": false 00:11:52.529 }, 00:11:52.529 "memory_domains": [ 00:11:52.529 { 00:11:52.529 "dma_device_id": "system", 00:11:52.529 "dma_device_type": 1 00:11:52.529 }, 00:11:52.529 { 00:11:52.529 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:52.529 "dma_device_type": 2 00:11:52.529 } 00:11:52.529 ], 00:11:52.529 "driver_specific": {} 00:11:52.529 } 00:11:52.529 ] 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.529 [2024-11-20 15:59:50.740216] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:52.529 [2024-11-20 15:59:50.740365] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:52.529 [2024-11-20 15:59:50.740433] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:52.529 [2024-11-20 15:59:50.742294] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:52.529 [2024-11-20 15:59:50.742426] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.529 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.790 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:52.790 "name": "Existed_Raid", 00:11:52.790 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:52.790 "strip_size_kb": 0, 00:11:52.790 "state": "configuring", 00:11:52.790 "raid_level": "raid1", 00:11:52.790 "superblock": false, 00:11:52.790 "num_base_bdevs": 4, 00:11:52.790 "num_base_bdevs_discovered": 3, 00:11:52.790 "num_base_bdevs_operational": 4, 00:11:52.790 "base_bdevs_list": [ 00:11:52.790 { 00:11:52.790 "name": "BaseBdev1", 00:11:52.790 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:52.790 "is_configured": false, 00:11:52.790 "data_offset": 0, 00:11:52.790 "data_size": 0 00:11:52.790 }, 00:11:52.790 { 00:11:52.790 "name": "BaseBdev2", 00:11:52.790 "uuid": "7bedd1d4-b5db-4f3b-8e8c-7097104aae97", 00:11:52.790 "is_configured": true, 00:11:52.790 "data_offset": 0, 00:11:52.790 "data_size": 65536 00:11:52.790 }, 00:11:52.790 { 00:11:52.790 "name": "BaseBdev3", 00:11:52.790 "uuid": "b92061de-c461-41e2-b990-a9624a0c9d9a", 00:11:52.790 "is_configured": true, 00:11:52.790 "data_offset": 0, 00:11:52.790 "data_size": 65536 00:11:52.790 }, 00:11:52.790 { 00:11:52.790 "name": "BaseBdev4", 00:11:52.790 "uuid": "0455bb9f-a2fa-40ca-8a67-5cd915a4f050", 00:11:52.790 "is_configured": true, 00:11:52.790 "data_offset": 0, 00:11:52.790 "data_size": 65536 00:11:52.790 } 00:11:52.790 ] 00:11:52.790 }' 00:11:52.790 15:59:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:52.790 15:59:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.052 [2024-11-20 15:59:51.056304] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:53.052 "name": "Existed_Raid", 00:11:53.052 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:53.052 "strip_size_kb": 0, 00:11:53.052 "state": "configuring", 00:11:53.052 "raid_level": "raid1", 00:11:53.052 "superblock": false, 00:11:53.052 "num_base_bdevs": 4, 00:11:53.052 "num_base_bdevs_discovered": 2, 00:11:53.052 "num_base_bdevs_operational": 4, 00:11:53.052 "base_bdevs_list": [ 00:11:53.052 { 00:11:53.052 "name": "BaseBdev1", 00:11:53.052 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:53.052 "is_configured": false, 00:11:53.052 "data_offset": 0, 00:11:53.052 "data_size": 0 00:11:53.052 }, 00:11:53.052 { 00:11:53.052 "name": null, 00:11:53.052 "uuid": "7bedd1d4-b5db-4f3b-8e8c-7097104aae97", 00:11:53.052 "is_configured": false, 00:11:53.052 "data_offset": 0, 00:11:53.052 "data_size": 65536 00:11:53.052 }, 00:11:53.052 { 00:11:53.052 "name": "BaseBdev3", 00:11:53.052 "uuid": "b92061de-c461-41e2-b990-a9624a0c9d9a", 00:11:53.052 "is_configured": true, 00:11:53.052 "data_offset": 0, 00:11:53.052 "data_size": 65536 00:11:53.052 }, 00:11:53.052 { 00:11:53.052 "name": "BaseBdev4", 00:11:53.052 "uuid": "0455bb9f-a2fa-40ca-8a67-5cd915a4f050", 00:11:53.052 "is_configured": true, 00:11:53.052 "data_offset": 0, 00:11:53.052 "data_size": 65536 00:11:53.052 } 00:11:53.052 ] 00:11:53.052 }' 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:53.052 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.315 [2024-11-20 15:59:51.450834] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:53.315 BaseBdev1 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.315 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.316 [ 00:11:53.316 { 00:11:53.316 "name": "BaseBdev1", 00:11:53.316 "aliases": [ 00:11:53.316 "ee5fb32f-0c3a-437b-9556-89c47d9ed429" 00:11:53.316 ], 00:11:53.316 "product_name": "Malloc disk", 00:11:53.316 "block_size": 512, 00:11:53.316 "num_blocks": 65536, 00:11:53.316 "uuid": "ee5fb32f-0c3a-437b-9556-89c47d9ed429", 00:11:53.316 "assigned_rate_limits": { 00:11:53.316 "rw_ios_per_sec": 0, 00:11:53.316 "rw_mbytes_per_sec": 0, 00:11:53.316 "r_mbytes_per_sec": 0, 00:11:53.316 "w_mbytes_per_sec": 0 00:11:53.316 }, 00:11:53.316 "claimed": true, 00:11:53.316 "claim_type": "exclusive_write", 00:11:53.316 "zoned": false, 00:11:53.316 "supported_io_types": { 00:11:53.316 "read": true, 00:11:53.316 "write": true, 00:11:53.316 "unmap": true, 00:11:53.316 "flush": true, 00:11:53.316 "reset": true, 00:11:53.316 "nvme_admin": false, 00:11:53.316 "nvme_io": false, 00:11:53.316 "nvme_io_md": false, 00:11:53.316 "write_zeroes": true, 00:11:53.316 "zcopy": true, 00:11:53.316 "get_zone_info": false, 00:11:53.316 "zone_management": false, 00:11:53.316 "zone_append": false, 00:11:53.316 "compare": false, 00:11:53.316 "compare_and_write": false, 00:11:53.316 "abort": true, 00:11:53.316 "seek_hole": false, 00:11:53.316 "seek_data": false, 00:11:53.316 "copy": true, 00:11:53.316 "nvme_iov_md": false 00:11:53.316 }, 00:11:53.316 "memory_domains": [ 00:11:53.316 { 00:11:53.316 "dma_device_id": "system", 00:11:53.316 "dma_device_type": 1 00:11:53.316 }, 00:11:53.316 { 00:11:53.316 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:53.316 "dma_device_type": 2 00:11:53.316 } 00:11:53.316 ], 00:11:53.316 "driver_specific": {} 00:11:53.316 } 00:11:53.316 ] 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:53.316 "name": "Existed_Raid", 00:11:53.316 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:53.316 "strip_size_kb": 0, 00:11:53.316 "state": "configuring", 00:11:53.316 "raid_level": "raid1", 00:11:53.316 "superblock": false, 00:11:53.316 "num_base_bdevs": 4, 00:11:53.316 "num_base_bdevs_discovered": 3, 00:11:53.316 "num_base_bdevs_operational": 4, 00:11:53.316 "base_bdevs_list": [ 00:11:53.316 { 00:11:53.316 "name": "BaseBdev1", 00:11:53.316 "uuid": "ee5fb32f-0c3a-437b-9556-89c47d9ed429", 00:11:53.316 "is_configured": true, 00:11:53.316 "data_offset": 0, 00:11:53.316 "data_size": 65536 00:11:53.316 }, 00:11:53.316 { 00:11:53.316 "name": null, 00:11:53.316 "uuid": "7bedd1d4-b5db-4f3b-8e8c-7097104aae97", 00:11:53.316 "is_configured": false, 00:11:53.316 "data_offset": 0, 00:11:53.316 "data_size": 65536 00:11:53.316 }, 00:11:53.316 { 00:11:53.316 "name": "BaseBdev3", 00:11:53.316 "uuid": "b92061de-c461-41e2-b990-a9624a0c9d9a", 00:11:53.316 "is_configured": true, 00:11:53.316 "data_offset": 0, 00:11:53.316 "data_size": 65536 00:11:53.316 }, 00:11:53.316 { 00:11:53.316 "name": "BaseBdev4", 00:11:53.316 "uuid": "0455bb9f-a2fa-40ca-8a67-5cd915a4f050", 00:11:53.316 "is_configured": true, 00:11:53.316 "data_offset": 0, 00:11:53.316 "data_size": 65536 00:11:53.316 } 00:11:53.316 ] 00:11:53.316 }' 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:53.316 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.577 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:53.577 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.577 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.577 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.577 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.838 [2024-11-20 15:59:51.846997] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:53.838 "name": "Existed_Raid", 00:11:53.838 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:53.838 "strip_size_kb": 0, 00:11:53.838 "state": "configuring", 00:11:53.838 "raid_level": "raid1", 00:11:53.838 "superblock": false, 00:11:53.838 "num_base_bdevs": 4, 00:11:53.838 "num_base_bdevs_discovered": 2, 00:11:53.838 "num_base_bdevs_operational": 4, 00:11:53.838 "base_bdevs_list": [ 00:11:53.838 { 00:11:53.838 "name": "BaseBdev1", 00:11:53.838 "uuid": "ee5fb32f-0c3a-437b-9556-89c47d9ed429", 00:11:53.838 "is_configured": true, 00:11:53.838 "data_offset": 0, 00:11:53.838 "data_size": 65536 00:11:53.838 }, 00:11:53.838 { 00:11:53.838 "name": null, 00:11:53.838 "uuid": "7bedd1d4-b5db-4f3b-8e8c-7097104aae97", 00:11:53.838 "is_configured": false, 00:11:53.838 "data_offset": 0, 00:11:53.838 "data_size": 65536 00:11:53.838 }, 00:11:53.838 { 00:11:53.838 "name": null, 00:11:53.838 "uuid": "b92061de-c461-41e2-b990-a9624a0c9d9a", 00:11:53.838 "is_configured": false, 00:11:53.838 "data_offset": 0, 00:11:53.838 "data_size": 65536 00:11:53.838 }, 00:11:53.838 { 00:11:53.838 "name": "BaseBdev4", 00:11:53.838 "uuid": "0455bb9f-a2fa-40ca-8a67-5cd915a4f050", 00:11:53.838 "is_configured": true, 00:11:53.838 "data_offset": 0, 00:11:53.838 "data_size": 65536 00:11:53.838 } 00:11:53.838 ] 00:11:53.838 }' 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:53.838 15:59:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.099 [2024-11-20 15:59:52.191074] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:54.099 "name": "Existed_Raid", 00:11:54.099 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:54.099 "strip_size_kb": 0, 00:11:54.099 "state": "configuring", 00:11:54.099 "raid_level": "raid1", 00:11:54.099 "superblock": false, 00:11:54.099 "num_base_bdevs": 4, 00:11:54.099 "num_base_bdevs_discovered": 3, 00:11:54.099 "num_base_bdevs_operational": 4, 00:11:54.099 "base_bdevs_list": [ 00:11:54.099 { 00:11:54.099 "name": "BaseBdev1", 00:11:54.099 "uuid": "ee5fb32f-0c3a-437b-9556-89c47d9ed429", 00:11:54.099 "is_configured": true, 00:11:54.099 "data_offset": 0, 00:11:54.099 "data_size": 65536 00:11:54.099 }, 00:11:54.099 { 00:11:54.099 "name": null, 00:11:54.099 "uuid": "7bedd1d4-b5db-4f3b-8e8c-7097104aae97", 00:11:54.099 "is_configured": false, 00:11:54.099 "data_offset": 0, 00:11:54.099 "data_size": 65536 00:11:54.099 }, 00:11:54.099 { 00:11:54.099 "name": "BaseBdev3", 00:11:54.099 "uuid": "b92061de-c461-41e2-b990-a9624a0c9d9a", 00:11:54.099 "is_configured": true, 00:11:54.099 "data_offset": 0, 00:11:54.099 "data_size": 65536 00:11:54.099 }, 00:11:54.099 { 00:11:54.099 "name": "BaseBdev4", 00:11:54.099 "uuid": "0455bb9f-a2fa-40ca-8a67-5cd915a4f050", 00:11:54.099 "is_configured": true, 00:11:54.099 "data_offset": 0, 00:11:54.099 "data_size": 65536 00:11:54.099 } 00:11:54.099 ] 00:11:54.099 }' 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:54.099 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.359 [2024-11-20 15:59:52.531202] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:54.359 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.618 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.618 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:54.618 "name": "Existed_Raid", 00:11:54.618 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:54.618 "strip_size_kb": 0, 00:11:54.618 "state": "configuring", 00:11:54.618 "raid_level": "raid1", 00:11:54.618 "superblock": false, 00:11:54.618 "num_base_bdevs": 4, 00:11:54.618 "num_base_bdevs_discovered": 2, 00:11:54.618 "num_base_bdevs_operational": 4, 00:11:54.618 "base_bdevs_list": [ 00:11:54.618 { 00:11:54.618 "name": null, 00:11:54.618 "uuid": "ee5fb32f-0c3a-437b-9556-89c47d9ed429", 00:11:54.618 "is_configured": false, 00:11:54.618 "data_offset": 0, 00:11:54.618 "data_size": 65536 00:11:54.618 }, 00:11:54.618 { 00:11:54.618 "name": null, 00:11:54.618 "uuid": "7bedd1d4-b5db-4f3b-8e8c-7097104aae97", 00:11:54.618 "is_configured": false, 00:11:54.618 "data_offset": 0, 00:11:54.618 "data_size": 65536 00:11:54.618 }, 00:11:54.618 { 00:11:54.618 "name": "BaseBdev3", 00:11:54.618 "uuid": "b92061de-c461-41e2-b990-a9624a0c9d9a", 00:11:54.618 "is_configured": true, 00:11:54.618 "data_offset": 0, 00:11:54.618 "data_size": 65536 00:11:54.618 }, 00:11:54.618 { 00:11:54.618 "name": "BaseBdev4", 00:11:54.618 "uuid": "0455bb9f-a2fa-40ca-8a67-5cd915a4f050", 00:11:54.618 "is_configured": true, 00:11:54.618 "data_offset": 0, 00:11:54.618 "data_size": 65536 00:11:54.618 } 00:11:54.618 ] 00:11:54.618 }' 00:11:54.618 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:54.618 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.878 [2024-11-20 15:59:52.941842] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.878 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:54.878 "name": "Existed_Raid", 00:11:54.878 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:54.878 "strip_size_kb": 0, 00:11:54.878 "state": "configuring", 00:11:54.878 "raid_level": "raid1", 00:11:54.878 "superblock": false, 00:11:54.878 "num_base_bdevs": 4, 00:11:54.878 "num_base_bdevs_discovered": 3, 00:11:54.878 "num_base_bdevs_operational": 4, 00:11:54.878 "base_bdevs_list": [ 00:11:54.878 { 00:11:54.878 "name": null, 00:11:54.878 "uuid": "ee5fb32f-0c3a-437b-9556-89c47d9ed429", 00:11:54.878 "is_configured": false, 00:11:54.878 "data_offset": 0, 00:11:54.878 "data_size": 65536 00:11:54.878 }, 00:11:54.878 { 00:11:54.878 "name": "BaseBdev2", 00:11:54.878 "uuid": "7bedd1d4-b5db-4f3b-8e8c-7097104aae97", 00:11:54.878 "is_configured": true, 00:11:54.878 "data_offset": 0, 00:11:54.878 "data_size": 65536 00:11:54.878 }, 00:11:54.878 { 00:11:54.878 "name": "BaseBdev3", 00:11:54.878 "uuid": "b92061de-c461-41e2-b990-a9624a0c9d9a", 00:11:54.878 "is_configured": true, 00:11:54.878 "data_offset": 0, 00:11:54.878 "data_size": 65536 00:11:54.878 }, 00:11:54.878 { 00:11:54.878 "name": "BaseBdev4", 00:11:54.878 "uuid": "0455bb9f-a2fa-40ca-8a67-5cd915a4f050", 00:11:54.878 "is_configured": true, 00:11:54.878 "data_offset": 0, 00:11:54.879 "data_size": 65536 00:11:54.879 } 00:11:54.879 ] 00:11:54.879 }' 00:11:54.879 15:59:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:54.879 15:59:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u ee5fb32f-0c3a-437b-9556-89c47d9ed429 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.140 [2024-11-20 15:59:53.348473] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:11:55.140 [2024-11-20 15:59:53.348516] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:11:55.140 [2024-11-20 15:59:53.348525] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:55.140 [2024-11-20 15:59:53.348800] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:11:55.140 [2024-11-20 15:59:53.348939] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:11:55.140 [2024-11-20 15:59:53.348948] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:11:55.140 [2024-11-20 15:59:53.349167] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:55.140 NewBaseBdev 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.140 [ 00:11:55.140 { 00:11:55.140 "name": "NewBaseBdev", 00:11:55.140 "aliases": [ 00:11:55.140 "ee5fb32f-0c3a-437b-9556-89c47d9ed429" 00:11:55.140 ], 00:11:55.140 "product_name": "Malloc disk", 00:11:55.140 "block_size": 512, 00:11:55.140 "num_blocks": 65536, 00:11:55.140 "uuid": "ee5fb32f-0c3a-437b-9556-89c47d9ed429", 00:11:55.140 "assigned_rate_limits": { 00:11:55.140 "rw_ios_per_sec": 0, 00:11:55.140 "rw_mbytes_per_sec": 0, 00:11:55.140 "r_mbytes_per_sec": 0, 00:11:55.140 "w_mbytes_per_sec": 0 00:11:55.140 }, 00:11:55.140 "claimed": true, 00:11:55.140 "claim_type": "exclusive_write", 00:11:55.140 "zoned": false, 00:11:55.140 "supported_io_types": { 00:11:55.140 "read": true, 00:11:55.140 "write": true, 00:11:55.140 "unmap": true, 00:11:55.140 "flush": true, 00:11:55.140 "reset": true, 00:11:55.140 "nvme_admin": false, 00:11:55.140 "nvme_io": false, 00:11:55.140 "nvme_io_md": false, 00:11:55.140 "write_zeroes": true, 00:11:55.140 "zcopy": true, 00:11:55.140 "get_zone_info": false, 00:11:55.140 "zone_management": false, 00:11:55.140 "zone_append": false, 00:11:55.140 "compare": false, 00:11:55.140 "compare_and_write": false, 00:11:55.140 "abort": true, 00:11:55.140 "seek_hole": false, 00:11:55.140 "seek_data": false, 00:11:55.140 "copy": true, 00:11:55.140 "nvme_iov_md": false 00:11:55.140 }, 00:11:55.140 "memory_domains": [ 00:11:55.140 { 00:11:55.140 "dma_device_id": "system", 00:11:55.140 "dma_device_type": 1 00:11:55.140 }, 00:11:55.140 { 00:11:55.140 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:55.140 "dma_device_type": 2 00:11:55.140 } 00:11:55.140 ], 00:11:55.140 "driver_specific": {} 00:11:55.140 } 00:11:55.140 ] 00:11:55.140 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.141 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:55.141 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:11:55.141 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:55.141 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:55.141 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:55.141 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:55.141 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:55.141 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:55.141 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:55.141 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:55.141 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:55.141 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.141 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:55.141 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.141 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.401 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.401 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:55.401 "name": "Existed_Raid", 00:11:55.401 "uuid": "f1485a32-120e-4673-8ab0-15921760112d", 00:11:55.401 "strip_size_kb": 0, 00:11:55.401 "state": "online", 00:11:55.401 "raid_level": "raid1", 00:11:55.401 "superblock": false, 00:11:55.401 "num_base_bdevs": 4, 00:11:55.401 "num_base_bdevs_discovered": 4, 00:11:55.401 "num_base_bdevs_operational": 4, 00:11:55.401 "base_bdevs_list": [ 00:11:55.401 { 00:11:55.401 "name": "NewBaseBdev", 00:11:55.401 "uuid": "ee5fb32f-0c3a-437b-9556-89c47d9ed429", 00:11:55.401 "is_configured": true, 00:11:55.401 "data_offset": 0, 00:11:55.401 "data_size": 65536 00:11:55.401 }, 00:11:55.401 { 00:11:55.401 "name": "BaseBdev2", 00:11:55.401 "uuid": "7bedd1d4-b5db-4f3b-8e8c-7097104aae97", 00:11:55.401 "is_configured": true, 00:11:55.401 "data_offset": 0, 00:11:55.401 "data_size": 65536 00:11:55.401 }, 00:11:55.401 { 00:11:55.401 "name": "BaseBdev3", 00:11:55.401 "uuid": "b92061de-c461-41e2-b990-a9624a0c9d9a", 00:11:55.401 "is_configured": true, 00:11:55.401 "data_offset": 0, 00:11:55.401 "data_size": 65536 00:11:55.401 }, 00:11:55.401 { 00:11:55.401 "name": "BaseBdev4", 00:11:55.401 "uuid": "0455bb9f-a2fa-40ca-8a67-5cd915a4f050", 00:11:55.401 "is_configured": true, 00:11:55.401 "data_offset": 0, 00:11:55.401 "data_size": 65536 00:11:55.401 } 00:11:55.401 ] 00:11:55.401 }' 00:11:55.401 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:55.401 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.661 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:11:55.661 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:55.661 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:55.661 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:55.661 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:55.661 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:55.661 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:55.661 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:55.661 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.661 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.661 [2024-11-20 15:59:53.680975] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:55.661 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.661 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:55.661 "name": "Existed_Raid", 00:11:55.661 "aliases": [ 00:11:55.661 "f1485a32-120e-4673-8ab0-15921760112d" 00:11:55.661 ], 00:11:55.661 "product_name": "Raid Volume", 00:11:55.661 "block_size": 512, 00:11:55.661 "num_blocks": 65536, 00:11:55.661 "uuid": "f1485a32-120e-4673-8ab0-15921760112d", 00:11:55.661 "assigned_rate_limits": { 00:11:55.661 "rw_ios_per_sec": 0, 00:11:55.661 "rw_mbytes_per_sec": 0, 00:11:55.661 "r_mbytes_per_sec": 0, 00:11:55.661 "w_mbytes_per_sec": 0 00:11:55.661 }, 00:11:55.661 "claimed": false, 00:11:55.661 "zoned": false, 00:11:55.661 "supported_io_types": { 00:11:55.661 "read": true, 00:11:55.661 "write": true, 00:11:55.661 "unmap": false, 00:11:55.661 "flush": false, 00:11:55.661 "reset": true, 00:11:55.661 "nvme_admin": false, 00:11:55.661 "nvme_io": false, 00:11:55.661 "nvme_io_md": false, 00:11:55.661 "write_zeroes": true, 00:11:55.661 "zcopy": false, 00:11:55.661 "get_zone_info": false, 00:11:55.661 "zone_management": false, 00:11:55.661 "zone_append": false, 00:11:55.661 "compare": false, 00:11:55.661 "compare_and_write": false, 00:11:55.661 "abort": false, 00:11:55.661 "seek_hole": false, 00:11:55.661 "seek_data": false, 00:11:55.661 "copy": false, 00:11:55.661 "nvme_iov_md": false 00:11:55.661 }, 00:11:55.661 "memory_domains": [ 00:11:55.661 { 00:11:55.661 "dma_device_id": "system", 00:11:55.661 "dma_device_type": 1 00:11:55.661 }, 00:11:55.661 { 00:11:55.661 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:55.661 "dma_device_type": 2 00:11:55.661 }, 00:11:55.661 { 00:11:55.661 "dma_device_id": "system", 00:11:55.661 "dma_device_type": 1 00:11:55.661 }, 00:11:55.661 { 00:11:55.662 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:55.662 "dma_device_type": 2 00:11:55.662 }, 00:11:55.662 { 00:11:55.662 "dma_device_id": "system", 00:11:55.662 "dma_device_type": 1 00:11:55.662 }, 00:11:55.662 { 00:11:55.662 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:55.662 "dma_device_type": 2 00:11:55.662 }, 00:11:55.662 { 00:11:55.662 "dma_device_id": "system", 00:11:55.662 "dma_device_type": 1 00:11:55.662 }, 00:11:55.662 { 00:11:55.662 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:55.662 "dma_device_type": 2 00:11:55.662 } 00:11:55.662 ], 00:11:55.662 "driver_specific": { 00:11:55.662 "raid": { 00:11:55.662 "uuid": "f1485a32-120e-4673-8ab0-15921760112d", 00:11:55.662 "strip_size_kb": 0, 00:11:55.662 "state": "online", 00:11:55.662 "raid_level": "raid1", 00:11:55.662 "superblock": false, 00:11:55.662 "num_base_bdevs": 4, 00:11:55.662 "num_base_bdevs_discovered": 4, 00:11:55.662 "num_base_bdevs_operational": 4, 00:11:55.662 "base_bdevs_list": [ 00:11:55.662 { 00:11:55.662 "name": "NewBaseBdev", 00:11:55.662 "uuid": "ee5fb32f-0c3a-437b-9556-89c47d9ed429", 00:11:55.662 "is_configured": true, 00:11:55.662 "data_offset": 0, 00:11:55.662 "data_size": 65536 00:11:55.662 }, 00:11:55.662 { 00:11:55.662 "name": "BaseBdev2", 00:11:55.662 "uuid": "7bedd1d4-b5db-4f3b-8e8c-7097104aae97", 00:11:55.662 "is_configured": true, 00:11:55.662 "data_offset": 0, 00:11:55.662 "data_size": 65536 00:11:55.662 }, 00:11:55.662 { 00:11:55.662 "name": "BaseBdev3", 00:11:55.662 "uuid": "b92061de-c461-41e2-b990-a9624a0c9d9a", 00:11:55.662 "is_configured": true, 00:11:55.662 "data_offset": 0, 00:11:55.662 "data_size": 65536 00:11:55.662 }, 00:11:55.662 { 00:11:55.662 "name": "BaseBdev4", 00:11:55.662 "uuid": "0455bb9f-a2fa-40ca-8a67-5cd915a4f050", 00:11:55.662 "is_configured": true, 00:11:55.662 "data_offset": 0, 00:11:55.662 "data_size": 65536 00:11:55.662 } 00:11:55.662 ] 00:11:55.662 } 00:11:55.662 } 00:11:55.662 }' 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:11:55.662 BaseBdev2 00:11:55.662 BaseBdev3 00:11:55.662 BaseBdev4' 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.662 [2024-11-20 15:59:53.888635] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:55.662 [2024-11-20 15:59:53.888765] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:55.662 [2024-11-20 15:59:53.888852] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:55.662 [2024-11-20 15:59:53.889134] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:55.662 [2024-11-20 15:59:53.889146] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 71155 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 71155 ']' 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 71155 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:55.662 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71155 00:11:55.922 killing process with pid 71155 00:11:55.922 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:55.922 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:55.922 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71155' 00:11:55.922 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 71155 00:11:55.922 [2024-11-20 15:59:53.916824] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:55.922 15:59:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 71155 00:11:55.922 [2024-11-20 15:59:54.167324] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:56.862 ************************************ 00:11:56.862 END TEST raid_state_function_test 00:11:56.862 ************************************ 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:11:56.862 00:11:56.862 real 0m8.380s 00:11:56.862 user 0m13.267s 00:11:56.862 sys 0m1.401s 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:56.862 15:59:54 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 4 true 00:11:56.862 15:59:54 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:56.862 15:59:54 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:56.862 15:59:54 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:56.862 ************************************ 00:11:56.862 START TEST raid_state_function_test_sb 00:11:56.862 ************************************ 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 4 true 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:56.862 Process raid pid: 71793 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=71793 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 71793' 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 71793 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 71793 ']' 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:56.862 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:56.862 15:59:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:56.862 [2024-11-20 15:59:55.040623] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:11:56.862 [2024-11-20 15:59:55.040943] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:57.123 [2024-11-20 15:59:55.208325] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:57.123 [2024-11-20 15:59:55.310807] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:57.381 [2024-11-20 15:59:55.448099] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:57.381 [2024-11-20 15:59:55.448133] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:57.642 15:59:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:57.642 15:59:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:11:57.642 15:59:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:57.642 15:59:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:57.642 15:59:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:57.642 [2024-11-20 15:59:55.888877] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:57.642 [2024-11-20 15:59:55.888928] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:57.642 [2024-11-20 15:59:55.888939] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:57.642 [2024-11-20 15:59:55.888962] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:57.642 [2024-11-20 15:59:55.888969] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:57.642 [2024-11-20 15:59:55.888979] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:57.642 [2024-11-20 15:59:55.888985] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:57.642 [2024-11-20 15:59:55.888995] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:57.902 15:59:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:57.902 15:59:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:57.902 15:59:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:57.902 15:59:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:57.902 15:59:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:57.902 15:59:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:57.902 15:59:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:57.902 15:59:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:57.903 15:59:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:57.903 15:59:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:57.903 15:59:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:57.903 15:59:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:57.903 15:59:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:57.903 15:59:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:57.903 15:59:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:57.903 15:59:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:57.903 15:59:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:57.903 "name": "Existed_Raid", 00:11:57.903 "uuid": "d51eb372-3701-4b39-a4a5-69656fb69535", 00:11:57.903 "strip_size_kb": 0, 00:11:57.903 "state": "configuring", 00:11:57.903 "raid_level": "raid1", 00:11:57.903 "superblock": true, 00:11:57.903 "num_base_bdevs": 4, 00:11:57.903 "num_base_bdevs_discovered": 0, 00:11:57.903 "num_base_bdevs_operational": 4, 00:11:57.903 "base_bdevs_list": [ 00:11:57.903 { 00:11:57.903 "name": "BaseBdev1", 00:11:57.903 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:57.903 "is_configured": false, 00:11:57.903 "data_offset": 0, 00:11:57.903 "data_size": 0 00:11:57.903 }, 00:11:57.903 { 00:11:57.903 "name": "BaseBdev2", 00:11:57.903 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:57.903 "is_configured": false, 00:11:57.903 "data_offset": 0, 00:11:57.903 "data_size": 0 00:11:57.903 }, 00:11:57.903 { 00:11:57.903 "name": "BaseBdev3", 00:11:57.903 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:57.903 "is_configured": false, 00:11:57.903 "data_offset": 0, 00:11:57.903 "data_size": 0 00:11:57.903 }, 00:11:57.903 { 00:11:57.903 "name": "BaseBdev4", 00:11:57.903 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:57.903 "is_configured": false, 00:11:57.903 "data_offset": 0, 00:11:57.903 "data_size": 0 00:11:57.903 } 00:11:57.903 ] 00:11:57.903 }' 00:11:57.903 15:59:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:57.903 15:59:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.164 [2024-11-20 15:59:56.196918] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:58.164 [2024-11-20 15:59:56.196963] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.164 [2024-11-20 15:59:56.204959] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:58.164 [2024-11-20 15:59:56.205011] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:58.164 [2024-11-20 15:59:56.205025] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:58.164 [2024-11-20 15:59:56.205039] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:58.164 [2024-11-20 15:59:56.205049] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:58.164 [2024-11-20 15:59:56.205063] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:58.164 [2024-11-20 15:59:56.205073] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:58.164 [2024-11-20 15:59:56.205087] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.164 [2024-11-20 15:59:56.250952] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:58.164 BaseBdev1 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.164 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.164 [ 00:11:58.164 { 00:11:58.164 "name": "BaseBdev1", 00:11:58.164 "aliases": [ 00:11:58.164 "2eacf789-1c26-4126-8916-2c1e5f0bb48f" 00:11:58.164 ], 00:11:58.164 "product_name": "Malloc disk", 00:11:58.164 "block_size": 512, 00:11:58.164 "num_blocks": 65536, 00:11:58.164 "uuid": "2eacf789-1c26-4126-8916-2c1e5f0bb48f", 00:11:58.164 "assigned_rate_limits": { 00:11:58.164 "rw_ios_per_sec": 0, 00:11:58.164 "rw_mbytes_per_sec": 0, 00:11:58.164 "r_mbytes_per_sec": 0, 00:11:58.164 "w_mbytes_per_sec": 0 00:11:58.164 }, 00:11:58.164 "claimed": true, 00:11:58.164 "claim_type": "exclusive_write", 00:11:58.164 "zoned": false, 00:11:58.164 "supported_io_types": { 00:11:58.164 "read": true, 00:11:58.164 "write": true, 00:11:58.164 "unmap": true, 00:11:58.164 "flush": true, 00:11:58.164 "reset": true, 00:11:58.164 "nvme_admin": false, 00:11:58.164 "nvme_io": false, 00:11:58.164 "nvme_io_md": false, 00:11:58.164 "write_zeroes": true, 00:11:58.164 "zcopy": true, 00:11:58.164 "get_zone_info": false, 00:11:58.164 "zone_management": false, 00:11:58.164 "zone_append": false, 00:11:58.164 "compare": false, 00:11:58.164 "compare_and_write": false, 00:11:58.164 "abort": true, 00:11:58.164 "seek_hole": false, 00:11:58.164 "seek_data": false, 00:11:58.164 "copy": true, 00:11:58.164 "nvme_iov_md": false 00:11:58.164 }, 00:11:58.164 "memory_domains": [ 00:11:58.164 { 00:11:58.164 "dma_device_id": "system", 00:11:58.164 "dma_device_type": 1 00:11:58.164 }, 00:11:58.164 { 00:11:58.164 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:58.164 "dma_device_type": 2 00:11:58.164 } 00:11:58.164 ], 00:11:58.164 "driver_specific": {} 00:11:58.164 } 00:11:58.164 ] 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:58.165 "name": "Existed_Raid", 00:11:58.165 "uuid": "8f24e572-2c1a-4d92-9649-9846cb985587", 00:11:58.165 "strip_size_kb": 0, 00:11:58.165 "state": "configuring", 00:11:58.165 "raid_level": "raid1", 00:11:58.165 "superblock": true, 00:11:58.165 "num_base_bdevs": 4, 00:11:58.165 "num_base_bdevs_discovered": 1, 00:11:58.165 "num_base_bdevs_operational": 4, 00:11:58.165 "base_bdevs_list": [ 00:11:58.165 { 00:11:58.165 "name": "BaseBdev1", 00:11:58.165 "uuid": "2eacf789-1c26-4126-8916-2c1e5f0bb48f", 00:11:58.165 "is_configured": true, 00:11:58.165 "data_offset": 2048, 00:11:58.165 "data_size": 63488 00:11:58.165 }, 00:11:58.165 { 00:11:58.165 "name": "BaseBdev2", 00:11:58.165 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:58.165 "is_configured": false, 00:11:58.165 "data_offset": 0, 00:11:58.165 "data_size": 0 00:11:58.165 }, 00:11:58.165 { 00:11:58.165 "name": "BaseBdev3", 00:11:58.165 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:58.165 "is_configured": false, 00:11:58.165 "data_offset": 0, 00:11:58.165 "data_size": 0 00:11:58.165 }, 00:11:58.165 { 00:11:58.165 "name": "BaseBdev4", 00:11:58.165 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:58.165 "is_configured": false, 00:11:58.165 "data_offset": 0, 00:11:58.165 "data_size": 0 00:11:58.165 } 00:11:58.165 ] 00:11:58.165 }' 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:58.165 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.426 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:58.426 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.426 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.426 [2024-11-20 15:59:56.631076] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:58.427 [2024-11-20 15:59:56.631226] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.427 [2024-11-20 15:59:56.639117] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:58.427 [2024-11-20 15:59:56.640975] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:58.427 [2024-11-20 15:59:56.641012] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:58.427 [2024-11-20 15:59:56.641022] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:58.427 [2024-11-20 15:59:56.641033] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:58.427 [2024-11-20 15:59:56.641040] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:58.427 [2024-11-20 15:59:56.641048] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:58.427 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.687 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:58.687 "name": "Existed_Raid", 00:11:58.687 "uuid": "467cdfe0-213a-4a8c-a110-8617eb558c41", 00:11:58.687 "strip_size_kb": 0, 00:11:58.687 "state": "configuring", 00:11:58.687 "raid_level": "raid1", 00:11:58.687 "superblock": true, 00:11:58.687 "num_base_bdevs": 4, 00:11:58.687 "num_base_bdevs_discovered": 1, 00:11:58.687 "num_base_bdevs_operational": 4, 00:11:58.687 "base_bdevs_list": [ 00:11:58.687 { 00:11:58.687 "name": "BaseBdev1", 00:11:58.687 "uuid": "2eacf789-1c26-4126-8916-2c1e5f0bb48f", 00:11:58.687 "is_configured": true, 00:11:58.687 "data_offset": 2048, 00:11:58.687 "data_size": 63488 00:11:58.687 }, 00:11:58.687 { 00:11:58.687 "name": "BaseBdev2", 00:11:58.687 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:58.687 "is_configured": false, 00:11:58.687 "data_offset": 0, 00:11:58.687 "data_size": 0 00:11:58.687 }, 00:11:58.687 { 00:11:58.687 "name": "BaseBdev3", 00:11:58.687 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:58.687 "is_configured": false, 00:11:58.687 "data_offset": 0, 00:11:58.687 "data_size": 0 00:11:58.687 }, 00:11:58.687 { 00:11:58.687 "name": "BaseBdev4", 00:11:58.687 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:58.687 "is_configured": false, 00:11:58.687 "data_offset": 0, 00:11:58.687 "data_size": 0 00:11:58.687 } 00:11:58.687 ] 00:11:58.687 }' 00:11:58.687 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:58.687 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.947 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:58.947 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.947 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.947 [2024-11-20 15:59:56.989884] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:58.947 BaseBdev2 00:11:58.947 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.947 15:59:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:11:58.947 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:11:58.947 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:58.947 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:58.947 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:58.947 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:58.947 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:58.947 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.947 15:59:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.947 [ 00:11:58.947 { 00:11:58.947 "name": "BaseBdev2", 00:11:58.947 "aliases": [ 00:11:58.947 "eeec27e2-3d16-4ac9-9da4-5d8445c16fc8" 00:11:58.947 ], 00:11:58.947 "product_name": "Malloc disk", 00:11:58.947 "block_size": 512, 00:11:58.947 "num_blocks": 65536, 00:11:58.947 "uuid": "eeec27e2-3d16-4ac9-9da4-5d8445c16fc8", 00:11:58.947 "assigned_rate_limits": { 00:11:58.947 "rw_ios_per_sec": 0, 00:11:58.947 "rw_mbytes_per_sec": 0, 00:11:58.947 "r_mbytes_per_sec": 0, 00:11:58.947 "w_mbytes_per_sec": 0 00:11:58.947 }, 00:11:58.947 "claimed": true, 00:11:58.947 "claim_type": "exclusive_write", 00:11:58.947 "zoned": false, 00:11:58.947 "supported_io_types": { 00:11:58.947 "read": true, 00:11:58.947 "write": true, 00:11:58.947 "unmap": true, 00:11:58.947 "flush": true, 00:11:58.947 "reset": true, 00:11:58.947 "nvme_admin": false, 00:11:58.947 "nvme_io": false, 00:11:58.947 "nvme_io_md": false, 00:11:58.947 "write_zeroes": true, 00:11:58.947 "zcopy": true, 00:11:58.947 "get_zone_info": false, 00:11:58.947 "zone_management": false, 00:11:58.947 "zone_append": false, 00:11:58.947 "compare": false, 00:11:58.947 "compare_and_write": false, 00:11:58.947 "abort": true, 00:11:58.947 "seek_hole": false, 00:11:58.947 "seek_data": false, 00:11:58.947 "copy": true, 00:11:58.947 "nvme_iov_md": false 00:11:58.947 }, 00:11:58.947 "memory_domains": [ 00:11:58.947 { 00:11:58.947 "dma_device_id": "system", 00:11:58.947 "dma_device_type": 1 00:11:58.947 }, 00:11:58.947 { 00:11:58.947 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:58.947 "dma_device_type": 2 00:11:58.947 } 00:11:58.947 ], 00:11:58.947 "driver_specific": {} 00:11:58.947 } 00:11:58.947 ] 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:58.947 "name": "Existed_Raid", 00:11:58.947 "uuid": "467cdfe0-213a-4a8c-a110-8617eb558c41", 00:11:58.947 "strip_size_kb": 0, 00:11:58.947 "state": "configuring", 00:11:58.947 "raid_level": "raid1", 00:11:58.947 "superblock": true, 00:11:58.947 "num_base_bdevs": 4, 00:11:58.947 "num_base_bdevs_discovered": 2, 00:11:58.947 "num_base_bdevs_operational": 4, 00:11:58.947 "base_bdevs_list": [ 00:11:58.947 { 00:11:58.947 "name": "BaseBdev1", 00:11:58.947 "uuid": "2eacf789-1c26-4126-8916-2c1e5f0bb48f", 00:11:58.947 "is_configured": true, 00:11:58.947 "data_offset": 2048, 00:11:58.947 "data_size": 63488 00:11:58.947 }, 00:11:58.947 { 00:11:58.947 "name": "BaseBdev2", 00:11:58.947 "uuid": "eeec27e2-3d16-4ac9-9da4-5d8445c16fc8", 00:11:58.947 "is_configured": true, 00:11:58.947 "data_offset": 2048, 00:11:58.947 "data_size": 63488 00:11:58.947 }, 00:11:58.947 { 00:11:58.947 "name": "BaseBdev3", 00:11:58.947 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:58.947 "is_configured": false, 00:11:58.947 "data_offset": 0, 00:11:58.947 "data_size": 0 00:11:58.947 }, 00:11:58.947 { 00:11:58.947 "name": "BaseBdev4", 00:11:58.947 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:58.947 "is_configured": false, 00:11:58.947 "data_offset": 0, 00:11:58.947 "data_size": 0 00:11:58.947 } 00:11:58.947 ] 00:11:58.947 }' 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:58.947 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.207 [2024-11-20 15:59:57.382064] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:59.207 BaseBdev3 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.207 [ 00:11:59.207 { 00:11:59.207 "name": "BaseBdev3", 00:11:59.207 "aliases": [ 00:11:59.207 "6a7bad34-ca7f-4b9f-a467-aae1efb922c3" 00:11:59.207 ], 00:11:59.207 "product_name": "Malloc disk", 00:11:59.207 "block_size": 512, 00:11:59.207 "num_blocks": 65536, 00:11:59.207 "uuid": "6a7bad34-ca7f-4b9f-a467-aae1efb922c3", 00:11:59.207 "assigned_rate_limits": { 00:11:59.207 "rw_ios_per_sec": 0, 00:11:59.207 "rw_mbytes_per_sec": 0, 00:11:59.207 "r_mbytes_per_sec": 0, 00:11:59.207 "w_mbytes_per_sec": 0 00:11:59.207 }, 00:11:59.207 "claimed": true, 00:11:59.207 "claim_type": "exclusive_write", 00:11:59.207 "zoned": false, 00:11:59.207 "supported_io_types": { 00:11:59.207 "read": true, 00:11:59.207 "write": true, 00:11:59.207 "unmap": true, 00:11:59.207 "flush": true, 00:11:59.207 "reset": true, 00:11:59.207 "nvme_admin": false, 00:11:59.207 "nvme_io": false, 00:11:59.207 "nvme_io_md": false, 00:11:59.207 "write_zeroes": true, 00:11:59.207 "zcopy": true, 00:11:59.207 "get_zone_info": false, 00:11:59.207 "zone_management": false, 00:11:59.207 "zone_append": false, 00:11:59.207 "compare": false, 00:11:59.207 "compare_and_write": false, 00:11:59.207 "abort": true, 00:11:59.207 "seek_hole": false, 00:11:59.207 "seek_data": false, 00:11:59.207 "copy": true, 00:11:59.207 "nvme_iov_md": false 00:11:59.207 }, 00:11:59.207 "memory_domains": [ 00:11:59.207 { 00:11:59.207 "dma_device_id": "system", 00:11:59.207 "dma_device_type": 1 00:11:59.207 }, 00:11:59.207 { 00:11:59.207 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:59.207 "dma_device_type": 2 00:11:59.207 } 00:11:59.207 ], 00:11:59.207 "driver_specific": {} 00:11:59.207 } 00:11:59.207 ] 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:59.207 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:59.208 "name": "Existed_Raid", 00:11:59.208 "uuid": "467cdfe0-213a-4a8c-a110-8617eb558c41", 00:11:59.208 "strip_size_kb": 0, 00:11:59.208 "state": "configuring", 00:11:59.208 "raid_level": "raid1", 00:11:59.208 "superblock": true, 00:11:59.208 "num_base_bdevs": 4, 00:11:59.208 "num_base_bdevs_discovered": 3, 00:11:59.208 "num_base_bdevs_operational": 4, 00:11:59.208 "base_bdevs_list": [ 00:11:59.208 { 00:11:59.208 "name": "BaseBdev1", 00:11:59.208 "uuid": "2eacf789-1c26-4126-8916-2c1e5f0bb48f", 00:11:59.208 "is_configured": true, 00:11:59.208 "data_offset": 2048, 00:11:59.208 "data_size": 63488 00:11:59.208 }, 00:11:59.208 { 00:11:59.208 "name": "BaseBdev2", 00:11:59.208 "uuid": "eeec27e2-3d16-4ac9-9da4-5d8445c16fc8", 00:11:59.208 "is_configured": true, 00:11:59.208 "data_offset": 2048, 00:11:59.208 "data_size": 63488 00:11:59.208 }, 00:11:59.208 { 00:11:59.208 "name": "BaseBdev3", 00:11:59.208 "uuid": "6a7bad34-ca7f-4b9f-a467-aae1efb922c3", 00:11:59.208 "is_configured": true, 00:11:59.208 "data_offset": 2048, 00:11:59.208 "data_size": 63488 00:11:59.208 }, 00:11:59.208 { 00:11:59.208 "name": "BaseBdev4", 00:11:59.208 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:59.208 "is_configured": false, 00:11:59.208 "data_offset": 0, 00:11:59.208 "data_size": 0 00:11:59.208 } 00:11:59.208 ] 00:11:59.208 }' 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:59.208 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.779 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:59.779 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.779 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.779 [2024-11-20 15:59:57.761014] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:59.779 [2024-11-20 15:59:57.761255] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:11:59.779 [2024-11-20 15:59:57.761268] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:59.779 [2024-11-20 15:59:57.761536] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:11:59.779 [2024-11-20 15:59:57.761689] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:11:59.779 [2024-11-20 15:59:57.761701] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:11:59.779 [2024-11-20 15:59:57.761834] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:59.779 BaseBdev4 00:11:59.779 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.779 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:11:59.779 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:11:59.779 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:59.779 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:59.779 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:59.779 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:59.779 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:59.779 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.779 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.779 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.779 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:59.779 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.779 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.779 [ 00:11:59.779 { 00:11:59.780 "name": "BaseBdev4", 00:11:59.780 "aliases": [ 00:11:59.780 "5fc79252-0021-436d-8bd3-5e641cfd3945" 00:11:59.780 ], 00:11:59.780 "product_name": "Malloc disk", 00:11:59.780 "block_size": 512, 00:11:59.780 "num_blocks": 65536, 00:11:59.780 "uuid": "5fc79252-0021-436d-8bd3-5e641cfd3945", 00:11:59.780 "assigned_rate_limits": { 00:11:59.780 "rw_ios_per_sec": 0, 00:11:59.780 "rw_mbytes_per_sec": 0, 00:11:59.780 "r_mbytes_per_sec": 0, 00:11:59.780 "w_mbytes_per_sec": 0 00:11:59.780 }, 00:11:59.780 "claimed": true, 00:11:59.780 "claim_type": "exclusive_write", 00:11:59.780 "zoned": false, 00:11:59.780 "supported_io_types": { 00:11:59.780 "read": true, 00:11:59.780 "write": true, 00:11:59.780 "unmap": true, 00:11:59.780 "flush": true, 00:11:59.780 "reset": true, 00:11:59.780 "nvme_admin": false, 00:11:59.780 "nvme_io": false, 00:11:59.780 "nvme_io_md": false, 00:11:59.780 "write_zeroes": true, 00:11:59.780 "zcopy": true, 00:11:59.780 "get_zone_info": false, 00:11:59.780 "zone_management": false, 00:11:59.780 "zone_append": false, 00:11:59.780 "compare": false, 00:11:59.780 "compare_and_write": false, 00:11:59.780 "abort": true, 00:11:59.780 "seek_hole": false, 00:11:59.780 "seek_data": false, 00:11:59.780 "copy": true, 00:11:59.780 "nvme_iov_md": false 00:11:59.780 }, 00:11:59.780 "memory_domains": [ 00:11:59.780 { 00:11:59.780 "dma_device_id": "system", 00:11:59.780 "dma_device_type": 1 00:11:59.780 }, 00:11:59.780 { 00:11:59.780 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:59.780 "dma_device_type": 2 00:11:59.780 } 00:11:59.780 ], 00:11:59.780 "driver_specific": {} 00:11:59.780 } 00:11:59.780 ] 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:59.780 "name": "Existed_Raid", 00:11:59.780 "uuid": "467cdfe0-213a-4a8c-a110-8617eb558c41", 00:11:59.780 "strip_size_kb": 0, 00:11:59.780 "state": "online", 00:11:59.780 "raid_level": "raid1", 00:11:59.780 "superblock": true, 00:11:59.780 "num_base_bdevs": 4, 00:11:59.780 "num_base_bdevs_discovered": 4, 00:11:59.780 "num_base_bdevs_operational": 4, 00:11:59.780 "base_bdevs_list": [ 00:11:59.780 { 00:11:59.780 "name": "BaseBdev1", 00:11:59.780 "uuid": "2eacf789-1c26-4126-8916-2c1e5f0bb48f", 00:11:59.780 "is_configured": true, 00:11:59.780 "data_offset": 2048, 00:11:59.780 "data_size": 63488 00:11:59.780 }, 00:11:59.780 { 00:11:59.780 "name": "BaseBdev2", 00:11:59.780 "uuid": "eeec27e2-3d16-4ac9-9da4-5d8445c16fc8", 00:11:59.780 "is_configured": true, 00:11:59.780 "data_offset": 2048, 00:11:59.780 "data_size": 63488 00:11:59.780 }, 00:11:59.780 { 00:11:59.780 "name": "BaseBdev3", 00:11:59.780 "uuid": "6a7bad34-ca7f-4b9f-a467-aae1efb922c3", 00:11:59.780 "is_configured": true, 00:11:59.780 "data_offset": 2048, 00:11:59.780 "data_size": 63488 00:11:59.780 }, 00:11:59.780 { 00:11:59.780 "name": "BaseBdev4", 00:11:59.780 "uuid": "5fc79252-0021-436d-8bd3-5e641cfd3945", 00:11:59.780 "is_configured": true, 00:11:59.780 "data_offset": 2048, 00:11:59.780 "data_size": 63488 00:11:59.780 } 00:11:59.780 ] 00:11:59.780 }' 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:59.780 15:59:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.040 [2024-11-20 15:59:58.129493] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:12:00.040 "name": "Existed_Raid", 00:12:00.040 "aliases": [ 00:12:00.040 "467cdfe0-213a-4a8c-a110-8617eb558c41" 00:12:00.040 ], 00:12:00.040 "product_name": "Raid Volume", 00:12:00.040 "block_size": 512, 00:12:00.040 "num_blocks": 63488, 00:12:00.040 "uuid": "467cdfe0-213a-4a8c-a110-8617eb558c41", 00:12:00.040 "assigned_rate_limits": { 00:12:00.040 "rw_ios_per_sec": 0, 00:12:00.040 "rw_mbytes_per_sec": 0, 00:12:00.040 "r_mbytes_per_sec": 0, 00:12:00.040 "w_mbytes_per_sec": 0 00:12:00.040 }, 00:12:00.040 "claimed": false, 00:12:00.040 "zoned": false, 00:12:00.040 "supported_io_types": { 00:12:00.040 "read": true, 00:12:00.040 "write": true, 00:12:00.040 "unmap": false, 00:12:00.040 "flush": false, 00:12:00.040 "reset": true, 00:12:00.040 "nvme_admin": false, 00:12:00.040 "nvme_io": false, 00:12:00.040 "nvme_io_md": false, 00:12:00.040 "write_zeroes": true, 00:12:00.040 "zcopy": false, 00:12:00.040 "get_zone_info": false, 00:12:00.040 "zone_management": false, 00:12:00.040 "zone_append": false, 00:12:00.040 "compare": false, 00:12:00.040 "compare_and_write": false, 00:12:00.040 "abort": false, 00:12:00.040 "seek_hole": false, 00:12:00.040 "seek_data": false, 00:12:00.040 "copy": false, 00:12:00.040 "nvme_iov_md": false 00:12:00.040 }, 00:12:00.040 "memory_domains": [ 00:12:00.040 { 00:12:00.040 "dma_device_id": "system", 00:12:00.040 "dma_device_type": 1 00:12:00.040 }, 00:12:00.040 { 00:12:00.040 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:00.040 "dma_device_type": 2 00:12:00.040 }, 00:12:00.040 { 00:12:00.040 "dma_device_id": "system", 00:12:00.040 "dma_device_type": 1 00:12:00.040 }, 00:12:00.040 { 00:12:00.040 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:00.040 "dma_device_type": 2 00:12:00.040 }, 00:12:00.040 { 00:12:00.040 "dma_device_id": "system", 00:12:00.040 "dma_device_type": 1 00:12:00.040 }, 00:12:00.040 { 00:12:00.040 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:00.040 "dma_device_type": 2 00:12:00.040 }, 00:12:00.040 { 00:12:00.040 "dma_device_id": "system", 00:12:00.040 "dma_device_type": 1 00:12:00.040 }, 00:12:00.040 { 00:12:00.040 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:00.040 "dma_device_type": 2 00:12:00.040 } 00:12:00.040 ], 00:12:00.040 "driver_specific": { 00:12:00.040 "raid": { 00:12:00.040 "uuid": "467cdfe0-213a-4a8c-a110-8617eb558c41", 00:12:00.040 "strip_size_kb": 0, 00:12:00.040 "state": "online", 00:12:00.040 "raid_level": "raid1", 00:12:00.040 "superblock": true, 00:12:00.040 "num_base_bdevs": 4, 00:12:00.040 "num_base_bdevs_discovered": 4, 00:12:00.040 "num_base_bdevs_operational": 4, 00:12:00.040 "base_bdevs_list": [ 00:12:00.040 { 00:12:00.040 "name": "BaseBdev1", 00:12:00.040 "uuid": "2eacf789-1c26-4126-8916-2c1e5f0bb48f", 00:12:00.040 "is_configured": true, 00:12:00.040 "data_offset": 2048, 00:12:00.040 "data_size": 63488 00:12:00.040 }, 00:12:00.040 { 00:12:00.040 "name": "BaseBdev2", 00:12:00.040 "uuid": "eeec27e2-3d16-4ac9-9da4-5d8445c16fc8", 00:12:00.040 "is_configured": true, 00:12:00.040 "data_offset": 2048, 00:12:00.040 "data_size": 63488 00:12:00.040 }, 00:12:00.040 { 00:12:00.040 "name": "BaseBdev3", 00:12:00.040 "uuid": "6a7bad34-ca7f-4b9f-a467-aae1efb922c3", 00:12:00.040 "is_configured": true, 00:12:00.040 "data_offset": 2048, 00:12:00.040 "data_size": 63488 00:12:00.040 }, 00:12:00.040 { 00:12:00.040 "name": "BaseBdev4", 00:12:00.040 "uuid": "5fc79252-0021-436d-8bd3-5e641cfd3945", 00:12:00.040 "is_configured": true, 00:12:00.040 "data_offset": 2048, 00:12:00.040 "data_size": 63488 00:12:00.040 } 00:12:00.040 ] 00:12:00.040 } 00:12:00.040 } 00:12:00.040 }' 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:12:00.040 BaseBdev2 00:12:00.040 BaseBdev3 00:12:00.040 BaseBdev4' 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:00.040 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:00.041 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:00.041 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:12:00.041 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.041 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.041 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:00.041 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:00.301 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.302 [2024-11-20 15:59:58.357233] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:00.302 "name": "Existed_Raid", 00:12:00.302 "uuid": "467cdfe0-213a-4a8c-a110-8617eb558c41", 00:12:00.302 "strip_size_kb": 0, 00:12:00.302 "state": "online", 00:12:00.302 "raid_level": "raid1", 00:12:00.302 "superblock": true, 00:12:00.302 "num_base_bdevs": 4, 00:12:00.302 "num_base_bdevs_discovered": 3, 00:12:00.302 "num_base_bdevs_operational": 3, 00:12:00.302 "base_bdevs_list": [ 00:12:00.302 { 00:12:00.302 "name": null, 00:12:00.302 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:00.302 "is_configured": false, 00:12:00.302 "data_offset": 0, 00:12:00.302 "data_size": 63488 00:12:00.302 }, 00:12:00.302 { 00:12:00.302 "name": "BaseBdev2", 00:12:00.302 "uuid": "eeec27e2-3d16-4ac9-9da4-5d8445c16fc8", 00:12:00.302 "is_configured": true, 00:12:00.302 "data_offset": 2048, 00:12:00.302 "data_size": 63488 00:12:00.302 }, 00:12:00.302 { 00:12:00.302 "name": "BaseBdev3", 00:12:00.302 "uuid": "6a7bad34-ca7f-4b9f-a467-aae1efb922c3", 00:12:00.302 "is_configured": true, 00:12:00.302 "data_offset": 2048, 00:12:00.302 "data_size": 63488 00:12:00.302 }, 00:12:00.302 { 00:12:00.302 "name": "BaseBdev4", 00:12:00.302 "uuid": "5fc79252-0021-436d-8bd3-5e641cfd3945", 00:12:00.302 "is_configured": true, 00:12:00.302 "data_offset": 2048, 00:12:00.302 "data_size": 63488 00:12:00.302 } 00:12:00.302 ] 00:12:00.302 }' 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:00.302 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.563 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:12:00.563 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:12:00.563 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.563 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.563 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.563 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:12:00.563 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.563 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:12:00.563 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:12:00.563 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:12:00.563 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.563 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.563 [2024-11-20 15:59:58.787986] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.824 [2024-11-20 15:59:58.885836] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.824 15:59:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.824 [2024-11-20 15:59:58.984409] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:12:00.824 [2024-11-20 15:59:58.984500] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:00.824 [2024-11-20 15:59:59.043920] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:00.824 [2024-11-20 15:59:59.043966] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:00.824 [2024-11-20 15:59:59.043977] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:12:00.824 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.824 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:12:00.824 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:12:00.824 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:12:00.824 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.824 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.824 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.824 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.087 BaseBdev2 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.087 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.087 [ 00:12:01.087 { 00:12:01.087 "name": "BaseBdev2", 00:12:01.087 "aliases": [ 00:12:01.087 "f4e4fb28-1d2d-4018-a6df-8e1989d25b31" 00:12:01.087 ], 00:12:01.087 "product_name": "Malloc disk", 00:12:01.087 "block_size": 512, 00:12:01.087 "num_blocks": 65536, 00:12:01.087 "uuid": "f4e4fb28-1d2d-4018-a6df-8e1989d25b31", 00:12:01.087 "assigned_rate_limits": { 00:12:01.087 "rw_ios_per_sec": 0, 00:12:01.087 "rw_mbytes_per_sec": 0, 00:12:01.087 "r_mbytes_per_sec": 0, 00:12:01.087 "w_mbytes_per_sec": 0 00:12:01.087 }, 00:12:01.087 "claimed": false, 00:12:01.087 "zoned": false, 00:12:01.087 "supported_io_types": { 00:12:01.087 "read": true, 00:12:01.087 "write": true, 00:12:01.087 "unmap": true, 00:12:01.087 "flush": true, 00:12:01.087 "reset": true, 00:12:01.087 "nvme_admin": false, 00:12:01.087 "nvme_io": false, 00:12:01.087 "nvme_io_md": false, 00:12:01.087 "write_zeroes": true, 00:12:01.087 "zcopy": true, 00:12:01.087 "get_zone_info": false, 00:12:01.087 "zone_management": false, 00:12:01.087 "zone_append": false, 00:12:01.087 "compare": false, 00:12:01.087 "compare_and_write": false, 00:12:01.087 "abort": true, 00:12:01.087 "seek_hole": false, 00:12:01.087 "seek_data": false, 00:12:01.088 "copy": true, 00:12:01.088 "nvme_iov_md": false 00:12:01.088 }, 00:12:01.088 "memory_domains": [ 00:12:01.088 { 00:12:01.088 "dma_device_id": "system", 00:12:01.088 "dma_device_type": 1 00:12:01.088 }, 00:12:01.088 { 00:12:01.088 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:01.088 "dma_device_type": 2 00:12:01.088 } 00:12:01.088 ], 00:12:01.088 "driver_specific": {} 00:12:01.088 } 00:12:01.088 ] 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.088 BaseBdev3 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.088 [ 00:12:01.088 { 00:12:01.088 "name": "BaseBdev3", 00:12:01.088 "aliases": [ 00:12:01.088 "4c5ee9bc-0b14-4d88-bac4-4f7f7fe432c0" 00:12:01.088 ], 00:12:01.088 "product_name": "Malloc disk", 00:12:01.088 "block_size": 512, 00:12:01.088 "num_blocks": 65536, 00:12:01.088 "uuid": "4c5ee9bc-0b14-4d88-bac4-4f7f7fe432c0", 00:12:01.088 "assigned_rate_limits": { 00:12:01.088 "rw_ios_per_sec": 0, 00:12:01.088 "rw_mbytes_per_sec": 0, 00:12:01.088 "r_mbytes_per_sec": 0, 00:12:01.088 "w_mbytes_per_sec": 0 00:12:01.088 }, 00:12:01.088 "claimed": false, 00:12:01.088 "zoned": false, 00:12:01.088 "supported_io_types": { 00:12:01.088 "read": true, 00:12:01.088 "write": true, 00:12:01.088 "unmap": true, 00:12:01.088 "flush": true, 00:12:01.088 "reset": true, 00:12:01.088 "nvme_admin": false, 00:12:01.088 "nvme_io": false, 00:12:01.088 "nvme_io_md": false, 00:12:01.088 "write_zeroes": true, 00:12:01.088 "zcopy": true, 00:12:01.088 "get_zone_info": false, 00:12:01.088 "zone_management": false, 00:12:01.088 "zone_append": false, 00:12:01.088 "compare": false, 00:12:01.088 "compare_and_write": false, 00:12:01.088 "abort": true, 00:12:01.088 "seek_hole": false, 00:12:01.088 "seek_data": false, 00:12:01.088 "copy": true, 00:12:01.088 "nvme_iov_md": false 00:12:01.088 }, 00:12:01.088 "memory_domains": [ 00:12:01.088 { 00:12:01.088 "dma_device_id": "system", 00:12:01.088 "dma_device_type": 1 00:12:01.088 }, 00:12:01.088 { 00:12:01.088 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:01.088 "dma_device_type": 2 00:12:01.088 } 00:12:01.088 ], 00:12:01.088 "driver_specific": {} 00:12:01.088 } 00:12:01.088 ] 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.088 BaseBdev4 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.088 [ 00:12:01.088 { 00:12:01.088 "name": "BaseBdev4", 00:12:01.088 "aliases": [ 00:12:01.088 "dda1e045-4feb-467d-a83e-9c58840facbb" 00:12:01.088 ], 00:12:01.088 "product_name": "Malloc disk", 00:12:01.088 "block_size": 512, 00:12:01.088 "num_blocks": 65536, 00:12:01.088 "uuid": "dda1e045-4feb-467d-a83e-9c58840facbb", 00:12:01.088 "assigned_rate_limits": { 00:12:01.088 "rw_ios_per_sec": 0, 00:12:01.088 "rw_mbytes_per_sec": 0, 00:12:01.088 "r_mbytes_per_sec": 0, 00:12:01.088 "w_mbytes_per_sec": 0 00:12:01.088 }, 00:12:01.088 "claimed": false, 00:12:01.088 "zoned": false, 00:12:01.088 "supported_io_types": { 00:12:01.088 "read": true, 00:12:01.088 "write": true, 00:12:01.088 "unmap": true, 00:12:01.088 "flush": true, 00:12:01.088 "reset": true, 00:12:01.088 "nvme_admin": false, 00:12:01.088 "nvme_io": false, 00:12:01.088 "nvme_io_md": false, 00:12:01.088 "write_zeroes": true, 00:12:01.088 "zcopy": true, 00:12:01.088 "get_zone_info": false, 00:12:01.088 "zone_management": false, 00:12:01.088 "zone_append": false, 00:12:01.088 "compare": false, 00:12:01.088 "compare_and_write": false, 00:12:01.088 "abort": true, 00:12:01.088 "seek_hole": false, 00:12:01.088 "seek_data": false, 00:12:01.088 "copy": true, 00:12:01.088 "nvme_iov_md": false 00:12:01.088 }, 00:12:01.088 "memory_domains": [ 00:12:01.088 { 00:12:01.088 "dma_device_id": "system", 00:12:01.088 "dma_device_type": 1 00:12:01.088 }, 00:12:01.088 { 00:12:01.088 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:01.088 "dma_device_type": 2 00:12:01.088 } 00:12:01.088 ], 00:12:01.088 "driver_specific": {} 00:12:01.088 } 00:12:01.088 ] 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.088 [2024-11-20 15:59:59.256038] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:12:01.088 [2024-11-20 15:59:59.256078] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:12:01.088 [2024-11-20 15:59:59.256096] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:01.088 [2024-11-20 15:59:59.257941] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:01.088 [2024-11-20 15:59:59.257990] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:01.088 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:01.089 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:01.089 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:01.089 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:01.089 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:01.089 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:01.089 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:01.089 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:01.089 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:01.089 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.089 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.089 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.089 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:01.089 "name": "Existed_Raid", 00:12:01.089 "uuid": "af57afad-c3fb-4cc7-9bc1-3087a96957da", 00:12:01.089 "strip_size_kb": 0, 00:12:01.089 "state": "configuring", 00:12:01.089 "raid_level": "raid1", 00:12:01.089 "superblock": true, 00:12:01.089 "num_base_bdevs": 4, 00:12:01.089 "num_base_bdevs_discovered": 3, 00:12:01.089 "num_base_bdevs_operational": 4, 00:12:01.089 "base_bdevs_list": [ 00:12:01.089 { 00:12:01.089 "name": "BaseBdev1", 00:12:01.089 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:01.089 "is_configured": false, 00:12:01.089 "data_offset": 0, 00:12:01.089 "data_size": 0 00:12:01.089 }, 00:12:01.089 { 00:12:01.089 "name": "BaseBdev2", 00:12:01.089 "uuid": "f4e4fb28-1d2d-4018-a6df-8e1989d25b31", 00:12:01.089 "is_configured": true, 00:12:01.089 "data_offset": 2048, 00:12:01.089 "data_size": 63488 00:12:01.089 }, 00:12:01.089 { 00:12:01.089 "name": "BaseBdev3", 00:12:01.089 "uuid": "4c5ee9bc-0b14-4d88-bac4-4f7f7fe432c0", 00:12:01.089 "is_configured": true, 00:12:01.089 "data_offset": 2048, 00:12:01.089 "data_size": 63488 00:12:01.089 }, 00:12:01.089 { 00:12:01.089 "name": "BaseBdev4", 00:12:01.089 "uuid": "dda1e045-4feb-467d-a83e-9c58840facbb", 00:12:01.089 "is_configured": true, 00:12:01.089 "data_offset": 2048, 00:12:01.089 "data_size": 63488 00:12:01.089 } 00:12:01.089 ] 00:12:01.089 }' 00:12:01.089 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:01.089 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.349 [2024-11-20 15:59:59.576108] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.349 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.611 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:01.611 "name": "Existed_Raid", 00:12:01.611 "uuid": "af57afad-c3fb-4cc7-9bc1-3087a96957da", 00:12:01.611 "strip_size_kb": 0, 00:12:01.611 "state": "configuring", 00:12:01.611 "raid_level": "raid1", 00:12:01.611 "superblock": true, 00:12:01.611 "num_base_bdevs": 4, 00:12:01.611 "num_base_bdevs_discovered": 2, 00:12:01.611 "num_base_bdevs_operational": 4, 00:12:01.611 "base_bdevs_list": [ 00:12:01.611 { 00:12:01.611 "name": "BaseBdev1", 00:12:01.611 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:01.611 "is_configured": false, 00:12:01.611 "data_offset": 0, 00:12:01.611 "data_size": 0 00:12:01.611 }, 00:12:01.611 { 00:12:01.611 "name": null, 00:12:01.611 "uuid": "f4e4fb28-1d2d-4018-a6df-8e1989d25b31", 00:12:01.611 "is_configured": false, 00:12:01.611 "data_offset": 0, 00:12:01.611 "data_size": 63488 00:12:01.611 }, 00:12:01.611 { 00:12:01.611 "name": "BaseBdev3", 00:12:01.611 "uuid": "4c5ee9bc-0b14-4d88-bac4-4f7f7fe432c0", 00:12:01.611 "is_configured": true, 00:12:01.611 "data_offset": 2048, 00:12:01.611 "data_size": 63488 00:12:01.611 }, 00:12:01.611 { 00:12:01.611 "name": "BaseBdev4", 00:12:01.611 "uuid": "dda1e045-4feb-467d-a83e-9c58840facbb", 00:12:01.611 "is_configured": true, 00:12:01.611 "data_offset": 2048, 00:12:01.611 "data_size": 63488 00:12:01.611 } 00:12:01.611 ] 00:12:01.611 }' 00:12:01.611 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:01.611 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.873 [2024-11-20 15:59:59.938659] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:01.873 BaseBdev1 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.873 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.873 [ 00:12:01.873 { 00:12:01.873 "name": "BaseBdev1", 00:12:01.873 "aliases": [ 00:12:01.873 "988ebb61-1f89-4758-8481-f8f467fb1b29" 00:12:01.873 ], 00:12:01.873 "product_name": "Malloc disk", 00:12:01.873 "block_size": 512, 00:12:01.873 "num_blocks": 65536, 00:12:01.873 "uuid": "988ebb61-1f89-4758-8481-f8f467fb1b29", 00:12:01.873 "assigned_rate_limits": { 00:12:01.873 "rw_ios_per_sec": 0, 00:12:01.873 "rw_mbytes_per_sec": 0, 00:12:01.873 "r_mbytes_per_sec": 0, 00:12:01.873 "w_mbytes_per_sec": 0 00:12:01.873 }, 00:12:01.873 "claimed": true, 00:12:01.873 "claim_type": "exclusive_write", 00:12:01.873 "zoned": false, 00:12:01.873 "supported_io_types": { 00:12:01.873 "read": true, 00:12:01.873 "write": true, 00:12:01.873 "unmap": true, 00:12:01.873 "flush": true, 00:12:01.873 "reset": true, 00:12:01.873 "nvme_admin": false, 00:12:01.873 "nvme_io": false, 00:12:01.873 "nvme_io_md": false, 00:12:01.873 "write_zeroes": true, 00:12:01.873 "zcopy": true, 00:12:01.873 "get_zone_info": false, 00:12:01.873 "zone_management": false, 00:12:01.873 "zone_append": false, 00:12:01.873 "compare": false, 00:12:01.873 "compare_and_write": false, 00:12:01.873 "abort": true, 00:12:01.873 "seek_hole": false, 00:12:01.873 "seek_data": false, 00:12:01.873 "copy": true, 00:12:01.873 "nvme_iov_md": false 00:12:01.873 }, 00:12:01.873 "memory_domains": [ 00:12:01.873 { 00:12:01.873 "dma_device_id": "system", 00:12:01.873 "dma_device_type": 1 00:12:01.873 }, 00:12:01.873 { 00:12:01.873 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:01.873 "dma_device_type": 2 00:12:01.873 } 00:12:01.873 ], 00:12:01.874 "driver_specific": {} 00:12:01.874 } 00:12:01.874 ] 00:12:01.874 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.874 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:01.874 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:12:01.874 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:01.874 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:01.874 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:01.874 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:01.874 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:01.874 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:01.874 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:01.874 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:01.874 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:01.874 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:01.874 15:59:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:01.874 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.874 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.874 15:59:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.874 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:01.874 "name": "Existed_Raid", 00:12:01.874 "uuid": "af57afad-c3fb-4cc7-9bc1-3087a96957da", 00:12:01.874 "strip_size_kb": 0, 00:12:01.874 "state": "configuring", 00:12:01.874 "raid_level": "raid1", 00:12:01.874 "superblock": true, 00:12:01.874 "num_base_bdevs": 4, 00:12:01.874 "num_base_bdevs_discovered": 3, 00:12:01.874 "num_base_bdevs_operational": 4, 00:12:01.874 "base_bdevs_list": [ 00:12:01.874 { 00:12:01.874 "name": "BaseBdev1", 00:12:01.874 "uuid": "988ebb61-1f89-4758-8481-f8f467fb1b29", 00:12:01.874 "is_configured": true, 00:12:01.874 "data_offset": 2048, 00:12:01.874 "data_size": 63488 00:12:01.874 }, 00:12:01.874 { 00:12:01.874 "name": null, 00:12:01.874 "uuid": "f4e4fb28-1d2d-4018-a6df-8e1989d25b31", 00:12:01.874 "is_configured": false, 00:12:01.874 "data_offset": 0, 00:12:01.874 "data_size": 63488 00:12:01.874 }, 00:12:01.874 { 00:12:01.874 "name": "BaseBdev3", 00:12:01.874 "uuid": "4c5ee9bc-0b14-4d88-bac4-4f7f7fe432c0", 00:12:01.874 "is_configured": true, 00:12:01.874 "data_offset": 2048, 00:12:01.874 "data_size": 63488 00:12:01.874 }, 00:12:01.874 { 00:12:01.874 "name": "BaseBdev4", 00:12:01.874 "uuid": "dda1e045-4feb-467d-a83e-9c58840facbb", 00:12:01.874 "is_configured": true, 00:12:01.874 "data_offset": 2048, 00:12:01.874 "data_size": 63488 00:12:01.874 } 00:12:01.874 ] 00:12:01.874 }' 00:12:01.874 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:01.874 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.134 [2024-11-20 16:00:00.342833] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.134 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.459 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:02.459 "name": "Existed_Raid", 00:12:02.459 "uuid": "af57afad-c3fb-4cc7-9bc1-3087a96957da", 00:12:02.459 "strip_size_kb": 0, 00:12:02.459 "state": "configuring", 00:12:02.459 "raid_level": "raid1", 00:12:02.459 "superblock": true, 00:12:02.459 "num_base_bdevs": 4, 00:12:02.459 "num_base_bdevs_discovered": 2, 00:12:02.459 "num_base_bdevs_operational": 4, 00:12:02.459 "base_bdevs_list": [ 00:12:02.459 { 00:12:02.459 "name": "BaseBdev1", 00:12:02.459 "uuid": "988ebb61-1f89-4758-8481-f8f467fb1b29", 00:12:02.459 "is_configured": true, 00:12:02.459 "data_offset": 2048, 00:12:02.459 "data_size": 63488 00:12:02.459 }, 00:12:02.459 { 00:12:02.459 "name": null, 00:12:02.459 "uuid": "f4e4fb28-1d2d-4018-a6df-8e1989d25b31", 00:12:02.459 "is_configured": false, 00:12:02.459 "data_offset": 0, 00:12:02.459 "data_size": 63488 00:12:02.459 }, 00:12:02.459 { 00:12:02.459 "name": null, 00:12:02.459 "uuid": "4c5ee9bc-0b14-4d88-bac4-4f7f7fe432c0", 00:12:02.459 "is_configured": false, 00:12:02.459 "data_offset": 0, 00:12:02.459 "data_size": 63488 00:12:02.459 }, 00:12:02.459 { 00:12:02.459 "name": "BaseBdev4", 00:12:02.459 "uuid": "dda1e045-4feb-467d-a83e-9c58840facbb", 00:12:02.459 "is_configured": true, 00:12:02.459 "data_offset": 2048, 00:12:02.459 "data_size": 63488 00:12:02.459 } 00:12:02.459 ] 00:12:02.459 }' 00:12:02.459 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:02.459 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.747 [2024-11-20 16:00:00.706914] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:02.747 "name": "Existed_Raid", 00:12:02.747 "uuid": "af57afad-c3fb-4cc7-9bc1-3087a96957da", 00:12:02.747 "strip_size_kb": 0, 00:12:02.747 "state": "configuring", 00:12:02.747 "raid_level": "raid1", 00:12:02.747 "superblock": true, 00:12:02.747 "num_base_bdevs": 4, 00:12:02.747 "num_base_bdevs_discovered": 3, 00:12:02.747 "num_base_bdevs_operational": 4, 00:12:02.747 "base_bdevs_list": [ 00:12:02.747 { 00:12:02.747 "name": "BaseBdev1", 00:12:02.747 "uuid": "988ebb61-1f89-4758-8481-f8f467fb1b29", 00:12:02.747 "is_configured": true, 00:12:02.747 "data_offset": 2048, 00:12:02.747 "data_size": 63488 00:12:02.747 }, 00:12:02.747 { 00:12:02.747 "name": null, 00:12:02.747 "uuid": "f4e4fb28-1d2d-4018-a6df-8e1989d25b31", 00:12:02.747 "is_configured": false, 00:12:02.747 "data_offset": 0, 00:12:02.747 "data_size": 63488 00:12:02.747 }, 00:12:02.747 { 00:12:02.747 "name": "BaseBdev3", 00:12:02.747 "uuid": "4c5ee9bc-0b14-4d88-bac4-4f7f7fe432c0", 00:12:02.747 "is_configured": true, 00:12:02.747 "data_offset": 2048, 00:12:02.747 "data_size": 63488 00:12:02.747 }, 00:12:02.747 { 00:12:02.747 "name": "BaseBdev4", 00:12:02.747 "uuid": "dda1e045-4feb-467d-a83e-9c58840facbb", 00:12:02.747 "is_configured": true, 00:12:02.747 "data_offset": 2048, 00:12:02.747 "data_size": 63488 00:12:02.747 } 00:12:02.747 ] 00:12:02.747 }' 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:02.747 16:00:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.006 [2024-11-20 16:00:01.075037] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:03.006 "name": "Existed_Raid", 00:12:03.006 "uuid": "af57afad-c3fb-4cc7-9bc1-3087a96957da", 00:12:03.006 "strip_size_kb": 0, 00:12:03.006 "state": "configuring", 00:12:03.006 "raid_level": "raid1", 00:12:03.006 "superblock": true, 00:12:03.006 "num_base_bdevs": 4, 00:12:03.006 "num_base_bdevs_discovered": 2, 00:12:03.006 "num_base_bdevs_operational": 4, 00:12:03.006 "base_bdevs_list": [ 00:12:03.006 { 00:12:03.006 "name": null, 00:12:03.006 "uuid": "988ebb61-1f89-4758-8481-f8f467fb1b29", 00:12:03.006 "is_configured": false, 00:12:03.006 "data_offset": 0, 00:12:03.006 "data_size": 63488 00:12:03.006 }, 00:12:03.006 { 00:12:03.006 "name": null, 00:12:03.006 "uuid": "f4e4fb28-1d2d-4018-a6df-8e1989d25b31", 00:12:03.006 "is_configured": false, 00:12:03.006 "data_offset": 0, 00:12:03.006 "data_size": 63488 00:12:03.006 }, 00:12:03.006 { 00:12:03.006 "name": "BaseBdev3", 00:12:03.006 "uuid": "4c5ee9bc-0b14-4d88-bac4-4f7f7fe432c0", 00:12:03.006 "is_configured": true, 00:12:03.006 "data_offset": 2048, 00:12:03.006 "data_size": 63488 00:12:03.006 }, 00:12:03.006 { 00:12:03.006 "name": "BaseBdev4", 00:12:03.006 "uuid": "dda1e045-4feb-467d-a83e-9c58840facbb", 00:12:03.006 "is_configured": true, 00:12:03.006 "data_offset": 2048, 00:12:03.006 "data_size": 63488 00:12:03.006 } 00:12:03.006 ] 00:12:03.006 }' 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:03.006 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.266 [2024-11-20 16:00:01.485331] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.266 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.525 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:03.525 "name": "Existed_Raid", 00:12:03.525 "uuid": "af57afad-c3fb-4cc7-9bc1-3087a96957da", 00:12:03.525 "strip_size_kb": 0, 00:12:03.525 "state": "configuring", 00:12:03.525 "raid_level": "raid1", 00:12:03.525 "superblock": true, 00:12:03.525 "num_base_bdevs": 4, 00:12:03.525 "num_base_bdevs_discovered": 3, 00:12:03.525 "num_base_bdevs_operational": 4, 00:12:03.525 "base_bdevs_list": [ 00:12:03.525 { 00:12:03.525 "name": null, 00:12:03.525 "uuid": "988ebb61-1f89-4758-8481-f8f467fb1b29", 00:12:03.525 "is_configured": false, 00:12:03.525 "data_offset": 0, 00:12:03.525 "data_size": 63488 00:12:03.525 }, 00:12:03.525 { 00:12:03.525 "name": "BaseBdev2", 00:12:03.525 "uuid": "f4e4fb28-1d2d-4018-a6df-8e1989d25b31", 00:12:03.525 "is_configured": true, 00:12:03.525 "data_offset": 2048, 00:12:03.525 "data_size": 63488 00:12:03.525 }, 00:12:03.525 { 00:12:03.525 "name": "BaseBdev3", 00:12:03.525 "uuid": "4c5ee9bc-0b14-4d88-bac4-4f7f7fe432c0", 00:12:03.525 "is_configured": true, 00:12:03.525 "data_offset": 2048, 00:12:03.525 "data_size": 63488 00:12:03.525 }, 00:12:03.525 { 00:12:03.525 "name": "BaseBdev4", 00:12:03.525 "uuid": "dda1e045-4feb-467d-a83e-9c58840facbb", 00:12:03.525 "is_configured": true, 00:12:03.525 "data_offset": 2048, 00:12:03.525 "data_size": 63488 00:12:03.525 } 00:12:03.525 ] 00:12:03.525 }' 00:12:03.525 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:03.525 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.785 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.785 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:12:03.785 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.785 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.785 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.785 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:12:03.785 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.785 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.785 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.785 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:12:03.785 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.785 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 988ebb61-1f89-4758-8481-f8f467fb1b29 00:12:03.785 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.785 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.785 [2024-11-20 16:00:01.875781] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:12:03.785 [2024-11-20 16:00:01.876097] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:12:03.785 [2024-11-20 16:00:01.876119] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:03.785 [2024-11-20 16:00:01.876383] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:12:03.785 [2024-11-20 16:00:01.876518] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:12:03.785 [2024-11-20 16:00:01.876526] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:12:03.785 [2024-11-20 16:00:01.876642] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:03.785 NewBaseBdev 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.786 [ 00:12:03.786 { 00:12:03.786 "name": "NewBaseBdev", 00:12:03.786 "aliases": [ 00:12:03.786 "988ebb61-1f89-4758-8481-f8f467fb1b29" 00:12:03.786 ], 00:12:03.786 "product_name": "Malloc disk", 00:12:03.786 "block_size": 512, 00:12:03.786 "num_blocks": 65536, 00:12:03.786 "uuid": "988ebb61-1f89-4758-8481-f8f467fb1b29", 00:12:03.786 "assigned_rate_limits": { 00:12:03.786 "rw_ios_per_sec": 0, 00:12:03.786 "rw_mbytes_per_sec": 0, 00:12:03.786 "r_mbytes_per_sec": 0, 00:12:03.786 "w_mbytes_per_sec": 0 00:12:03.786 }, 00:12:03.786 "claimed": true, 00:12:03.786 "claim_type": "exclusive_write", 00:12:03.786 "zoned": false, 00:12:03.786 "supported_io_types": { 00:12:03.786 "read": true, 00:12:03.786 "write": true, 00:12:03.786 "unmap": true, 00:12:03.786 "flush": true, 00:12:03.786 "reset": true, 00:12:03.786 "nvme_admin": false, 00:12:03.786 "nvme_io": false, 00:12:03.786 "nvme_io_md": false, 00:12:03.786 "write_zeroes": true, 00:12:03.786 "zcopy": true, 00:12:03.786 "get_zone_info": false, 00:12:03.786 "zone_management": false, 00:12:03.786 "zone_append": false, 00:12:03.786 "compare": false, 00:12:03.786 "compare_and_write": false, 00:12:03.786 "abort": true, 00:12:03.786 "seek_hole": false, 00:12:03.786 "seek_data": false, 00:12:03.786 "copy": true, 00:12:03.786 "nvme_iov_md": false 00:12:03.786 }, 00:12:03.786 "memory_domains": [ 00:12:03.786 { 00:12:03.786 "dma_device_id": "system", 00:12:03.786 "dma_device_type": 1 00:12:03.786 }, 00:12:03.786 { 00:12:03.786 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:03.786 "dma_device_type": 2 00:12:03.786 } 00:12:03.786 ], 00:12:03.786 "driver_specific": {} 00:12:03.786 } 00:12:03.786 ] 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:03.786 "name": "Existed_Raid", 00:12:03.786 "uuid": "af57afad-c3fb-4cc7-9bc1-3087a96957da", 00:12:03.786 "strip_size_kb": 0, 00:12:03.786 "state": "online", 00:12:03.786 "raid_level": "raid1", 00:12:03.786 "superblock": true, 00:12:03.786 "num_base_bdevs": 4, 00:12:03.786 "num_base_bdevs_discovered": 4, 00:12:03.786 "num_base_bdevs_operational": 4, 00:12:03.786 "base_bdevs_list": [ 00:12:03.786 { 00:12:03.786 "name": "NewBaseBdev", 00:12:03.786 "uuid": "988ebb61-1f89-4758-8481-f8f467fb1b29", 00:12:03.786 "is_configured": true, 00:12:03.786 "data_offset": 2048, 00:12:03.786 "data_size": 63488 00:12:03.786 }, 00:12:03.786 { 00:12:03.786 "name": "BaseBdev2", 00:12:03.786 "uuid": "f4e4fb28-1d2d-4018-a6df-8e1989d25b31", 00:12:03.786 "is_configured": true, 00:12:03.786 "data_offset": 2048, 00:12:03.786 "data_size": 63488 00:12:03.786 }, 00:12:03.786 { 00:12:03.786 "name": "BaseBdev3", 00:12:03.786 "uuid": "4c5ee9bc-0b14-4d88-bac4-4f7f7fe432c0", 00:12:03.786 "is_configured": true, 00:12:03.786 "data_offset": 2048, 00:12:03.786 "data_size": 63488 00:12:03.786 }, 00:12:03.786 { 00:12:03.786 "name": "BaseBdev4", 00:12:03.786 "uuid": "dda1e045-4feb-467d-a83e-9c58840facbb", 00:12:03.786 "is_configured": true, 00:12:03.786 "data_offset": 2048, 00:12:03.786 "data_size": 63488 00:12:03.786 } 00:12:03.786 ] 00:12:03.786 }' 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:03.786 16:00:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.045 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:12:04.045 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:12:04.045 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:12:04.045 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:12:04.045 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:12:04.045 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:12:04.045 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:12:04.045 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:12:04.045 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.045 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.045 [2024-11-20 16:00:02.232266] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:04.045 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.045 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:12:04.045 "name": "Existed_Raid", 00:12:04.045 "aliases": [ 00:12:04.045 "af57afad-c3fb-4cc7-9bc1-3087a96957da" 00:12:04.045 ], 00:12:04.045 "product_name": "Raid Volume", 00:12:04.045 "block_size": 512, 00:12:04.045 "num_blocks": 63488, 00:12:04.045 "uuid": "af57afad-c3fb-4cc7-9bc1-3087a96957da", 00:12:04.045 "assigned_rate_limits": { 00:12:04.045 "rw_ios_per_sec": 0, 00:12:04.045 "rw_mbytes_per_sec": 0, 00:12:04.045 "r_mbytes_per_sec": 0, 00:12:04.045 "w_mbytes_per_sec": 0 00:12:04.045 }, 00:12:04.045 "claimed": false, 00:12:04.045 "zoned": false, 00:12:04.045 "supported_io_types": { 00:12:04.045 "read": true, 00:12:04.045 "write": true, 00:12:04.045 "unmap": false, 00:12:04.045 "flush": false, 00:12:04.045 "reset": true, 00:12:04.045 "nvme_admin": false, 00:12:04.045 "nvme_io": false, 00:12:04.045 "nvme_io_md": false, 00:12:04.045 "write_zeroes": true, 00:12:04.045 "zcopy": false, 00:12:04.045 "get_zone_info": false, 00:12:04.045 "zone_management": false, 00:12:04.045 "zone_append": false, 00:12:04.045 "compare": false, 00:12:04.045 "compare_and_write": false, 00:12:04.045 "abort": false, 00:12:04.045 "seek_hole": false, 00:12:04.045 "seek_data": false, 00:12:04.045 "copy": false, 00:12:04.045 "nvme_iov_md": false 00:12:04.045 }, 00:12:04.045 "memory_domains": [ 00:12:04.045 { 00:12:04.045 "dma_device_id": "system", 00:12:04.045 "dma_device_type": 1 00:12:04.045 }, 00:12:04.045 { 00:12:04.045 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:04.045 "dma_device_type": 2 00:12:04.045 }, 00:12:04.045 { 00:12:04.045 "dma_device_id": "system", 00:12:04.045 "dma_device_type": 1 00:12:04.045 }, 00:12:04.045 { 00:12:04.045 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:04.045 "dma_device_type": 2 00:12:04.045 }, 00:12:04.045 { 00:12:04.045 "dma_device_id": "system", 00:12:04.045 "dma_device_type": 1 00:12:04.045 }, 00:12:04.045 { 00:12:04.045 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:04.045 "dma_device_type": 2 00:12:04.045 }, 00:12:04.045 { 00:12:04.046 "dma_device_id": "system", 00:12:04.046 "dma_device_type": 1 00:12:04.046 }, 00:12:04.046 { 00:12:04.046 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:04.046 "dma_device_type": 2 00:12:04.046 } 00:12:04.046 ], 00:12:04.046 "driver_specific": { 00:12:04.046 "raid": { 00:12:04.046 "uuid": "af57afad-c3fb-4cc7-9bc1-3087a96957da", 00:12:04.046 "strip_size_kb": 0, 00:12:04.046 "state": "online", 00:12:04.046 "raid_level": "raid1", 00:12:04.046 "superblock": true, 00:12:04.046 "num_base_bdevs": 4, 00:12:04.046 "num_base_bdevs_discovered": 4, 00:12:04.046 "num_base_bdevs_operational": 4, 00:12:04.046 "base_bdevs_list": [ 00:12:04.046 { 00:12:04.046 "name": "NewBaseBdev", 00:12:04.046 "uuid": "988ebb61-1f89-4758-8481-f8f467fb1b29", 00:12:04.046 "is_configured": true, 00:12:04.046 "data_offset": 2048, 00:12:04.046 "data_size": 63488 00:12:04.046 }, 00:12:04.046 { 00:12:04.046 "name": "BaseBdev2", 00:12:04.046 "uuid": "f4e4fb28-1d2d-4018-a6df-8e1989d25b31", 00:12:04.046 "is_configured": true, 00:12:04.046 "data_offset": 2048, 00:12:04.046 "data_size": 63488 00:12:04.046 }, 00:12:04.046 { 00:12:04.046 "name": "BaseBdev3", 00:12:04.046 "uuid": "4c5ee9bc-0b14-4d88-bac4-4f7f7fe432c0", 00:12:04.046 "is_configured": true, 00:12:04.046 "data_offset": 2048, 00:12:04.046 "data_size": 63488 00:12:04.046 }, 00:12:04.046 { 00:12:04.046 "name": "BaseBdev4", 00:12:04.046 "uuid": "dda1e045-4feb-467d-a83e-9c58840facbb", 00:12:04.046 "is_configured": true, 00:12:04.046 "data_offset": 2048, 00:12:04.046 "data_size": 63488 00:12:04.046 } 00:12:04.046 ] 00:12:04.046 } 00:12:04.046 } 00:12:04.046 }' 00:12:04.046 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:04.046 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:12:04.046 BaseBdev2 00:12:04.046 BaseBdev3 00:12:04.046 BaseBdev4' 00:12:04.046 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.305 [2024-11-20 16:00:02.455935] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:12:04.305 [2024-11-20 16:00:02.455959] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:04.305 [2024-11-20 16:00:02.456026] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:04.305 [2024-11-20 16:00:02.456305] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:04.305 [2024-11-20 16:00:02.456317] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 71793 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 71793 ']' 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 71793 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71793 00:12:04.305 killing process with pid 71793 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71793' 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 71793 00:12:04.305 [2024-11-20 16:00:02.486394] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:04.305 16:00:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 71793 00:12:04.565 [2024-11-20 16:00:02.733791] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:05.508 ************************************ 00:12:05.508 END TEST raid_state_function_test_sb 00:12:05.508 ************************************ 00:12:05.508 16:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:12:05.508 00:12:05.508 real 0m8.472s 00:12:05.508 user 0m13.492s 00:12:05.508 sys 0m1.356s 00:12:05.508 16:00:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:05.508 16:00:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.508 16:00:03 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 4 00:12:05.508 16:00:03 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:12:05.508 16:00:03 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:05.508 16:00:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:05.508 ************************************ 00:12:05.508 START TEST raid_superblock_test 00:12:05.508 ************************************ 00:12:05.508 16:00:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 4 00:12:05.508 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:12:05.508 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:12:05.508 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:12:05.508 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:12:05.508 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:12:05.508 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:12:05.508 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:12:05.508 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:12:05.508 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:12:05.508 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:12:05.508 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:12:05.508 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:12:05.508 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:12:05.508 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:12:05.509 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:12:05.509 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:05.509 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=72431 00:12:05.509 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 72431 00:12:05.509 16:00:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 72431 ']' 00:12:05.509 16:00:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:05.509 16:00:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:05.509 16:00:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:05.509 16:00:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:05.509 16:00:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:12:05.509 16:00:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:05.509 [2024-11-20 16:00:03.586843] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:12:05.509 [2024-11-20 16:00:03.587125] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72431 ] 00:12:05.509 [2024-11-20 16:00:03.745573] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:05.770 [2024-11-20 16:00:03.848000] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:05.770 [2024-11-20 16:00:03.985090] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:05.770 [2024-11-20 16:00:03.985310] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:06.429 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:06.429 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:12:06.429 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:12:06.429 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:12:06.429 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:12:06.429 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:12:06.429 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:12:06.429 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:12:06.429 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:12:06.429 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:12:06.429 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:12:06.429 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:06.430 malloc1 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:06.430 [2024-11-20 16:00:04.470424] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:12:06.430 [2024-11-20 16:00:04.470707] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:06.430 [2024-11-20 16:00:04.470742] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:06.430 [2024-11-20 16:00:04.470753] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:06.430 [2024-11-20 16:00:04.473143] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:06.430 [2024-11-20 16:00:04.473183] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:12:06.430 pt1 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:06.430 malloc2 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:06.430 [2024-11-20 16:00:04.512845] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:12:06.430 [2024-11-20 16:00:04.513115] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:06.430 [2024-11-20 16:00:04.513153] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:06.430 [2024-11-20 16:00:04.513163] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:06.430 [2024-11-20 16:00:04.515580] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:06.430 [2024-11-20 16:00:04.515620] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:12:06.430 pt2 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:06.430 malloc3 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:06.430 [2024-11-20 16:00:04.563903] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:12:06.430 [2024-11-20 16:00:04.563980] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:06.430 [2024-11-20 16:00:04.564004] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:06.430 [2024-11-20 16:00:04.564014] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:06.430 [2024-11-20 16:00:04.566410] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:06.430 [2024-11-20 16:00:04.566455] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:12:06.430 pt3 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:06.430 malloc4 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:06.430 [2024-11-20 16:00:04.602378] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:12:06.430 [2024-11-20 16:00:04.602603] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:06.430 [2024-11-20 16:00:04.602630] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:12:06.430 [2024-11-20 16:00:04.602639] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:06.430 [2024-11-20 16:00:04.604991] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:06.430 [2024-11-20 16:00:04.605027] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:12:06.430 pt4 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:06.430 [2024-11-20 16:00:04.610405] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:12:06.430 [2024-11-20 16:00:04.612368] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:06.430 [2024-11-20 16:00:04.612436] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:12:06.430 [2024-11-20 16:00:04.612500] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:12:06.430 [2024-11-20 16:00:04.612714] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:12:06.430 [2024-11-20 16:00:04.612729] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:06.430 [2024-11-20 16:00:04.613029] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:12:06.430 [2024-11-20 16:00:04.613199] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:12:06.430 [2024-11-20 16:00:04.613212] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:12:06.430 [2024-11-20 16:00:04.613368] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.430 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:06.431 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:06.431 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.431 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:06.431 "name": "raid_bdev1", 00:12:06.431 "uuid": "e1537747-274c-40d6-8bc8-e51190446046", 00:12:06.431 "strip_size_kb": 0, 00:12:06.431 "state": "online", 00:12:06.431 "raid_level": "raid1", 00:12:06.431 "superblock": true, 00:12:06.431 "num_base_bdevs": 4, 00:12:06.431 "num_base_bdevs_discovered": 4, 00:12:06.431 "num_base_bdevs_operational": 4, 00:12:06.431 "base_bdevs_list": [ 00:12:06.431 { 00:12:06.431 "name": "pt1", 00:12:06.431 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:06.431 "is_configured": true, 00:12:06.431 "data_offset": 2048, 00:12:06.431 "data_size": 63488 00:12:06.431 }, 00:12:06.431 { 00:12:06.431 "name": "pt2", 00:12:06.431 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:06.431 "is_configured": true, 00:12:06.431 "data_offset": 2048, 00:12:06.431 "data_size": 63488 00:12:06.431 }, 00:12:06.431 { 00:12:06.431 "name": "pt3", 00:12:06.431 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:06.431 "is_configured": true, 00:12:06.431 "data_offset": 2048, 00:12:06.431 "data_size": 63488 00:12:06.431 }, 00:12:06.431 { 00:12:06.431 "name": "pt4", 00:12:06.431 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:06.431 "is_configured": true, 00:12:06.431 "data_offset": 2048, 00:12:06.431 "data_size": 63488 00:12:06.431 } 00:12:06.431 ] 00:12:06.431 }' 00:12:06.431 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:06.431 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.004 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:12:07.004 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:12:07.004 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:12:07.004 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:12:07.004 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:12:07.004 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:12:07.004 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:07.004 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.004 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.004 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:12:07.004 [2024-11-20 16:00:04.958830] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:07.004 16:00:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.004 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:12:07.004 "name": "raid_bdev1", 00:12:07.004 "aliases": [ 00:12:07.004 "e1537747-274c-40d6-8bc8-e51190446046" 00:12:07.004 ], 00:12:07.004 "product_name": "Raid Volume", 00:12:07.004 "block_size": 512, 00:12:07.004 "num_blocks": 63488, 00:12:07.004 "uuid": "e1537747-274c-40d6-8bc8-e51190446046", 00:12:07.004 "assigned_rate_limits": { 00:12:07.004 "rw_ios_per_sec": 0, 00:12:07.004 "rw_mbytes_per_sec": 0, 00:12:07.004 "r_mbytes_per_sec": 0, 00:12:07.004 "w_mbytes_per_sec": 0 00:12:07.004 }, 00:12:07.004 "claimed": false, 00:12:07.004 "zoned": false, 00:12:07.004 "supported_io_types": { 00:12:07.004 "read": true, 00:12:07.004 "write": true, 00:12:07.004 "unmap": false, 00:12:07.004 "flush": false, 00:12:07.004 "reset": true, 00:12:07.004 "nvme_admin": false, 00:12:07.004 "nvme_io": false, 00:12:07.004 "nvme_io_md": false, 00:12:07.004 "write_zeroes": true, 00:12:07.004 "zcopy": false, 00:12:07.004 "get_zone_info": false, 00:12:07.004 "zone_management": false, 00:12:07.004 "zone_append": false, 00:12:07.004 "compare": false, 00:12:07.004 "compare_and_write": false, 00:12:07.004 "abort": false, 00:12:07.004 "seek_hole": false, 00:12:07.004 "seek_data": false, 00:12:07.004 "copy": false, 00:12:07.004 "nvme_iov_md": false 00:12:07.004 }, 00:12:07.004 "memory_domains": [ 00:12:07.004 { 00:12:07.004 "dma_device_id": "system", 00:12:07.004 "dma_device_type": 1 00:12:07.004 }, 00:12:07.004 { 00:12:07.004 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:07.004 "dma_device_type": 2 00:12:07.004 }, 00:12:07.004 { 00:12:07.004 "dma_device_id": "system", 00:12:07.004 "dma_device_type": 1 00:12:07.004 }, 00:12:07.004 { 00:12:07.004 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:07.004 "dma_device_type": 2 00:12:07.004 }, 00:12:07.004 { 00:12:07.004 "dma_device_id": "system", 00:12:07.004 "dma_device_type": 1 00:12:07.004 }, 00:12:07.004 { 00:12:07.004 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:07.004 "dma_device_type": 2 00:12:07.004 }, 00:12:07.005 { 00:12:07.005 "dma_device_id": "system", 00:12:07.005 "dma_device_type": 1 00:12:07.005 }, 00:12:07.005 { 00:12:07.005 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:07.005 "dma_device_type": 2 00:12:07.005 } 00:12:07.005 ], 00:12:07.005 "driver_specific": { 00:12:07.005 "raid": { 00:12:07.005 "uuid": "e1537747-274c-40d6-8bc8-e51190446046", 00:12:07.005 "strip_size_kb": 0, 00:12:07.005 "state": "online", 00:12:07.005 "raid_level": "raid1", 00:12:07.005 "superblock": true, 00:12:07.005 "num_base_bdevs": 4, 00:12:07.005 "num_base_bdevs_discovered": 4, 00:12:07.005 "num_base_bdevs_operational": 4, 00:12:07.005 "base_bdevs_list": [ 00:12:07.005 { 00:12:07.005 "name": "pt1", 00:12:07.005 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:07.005 "is_configured": true, 00:12:07.005 "data_offset": 2048, 00:12:07.005 "data_size": 63488 00:12:07.005 }, 00:12:07.005 { 00:12:07.005 "name": "pt2", 00:12:07.005 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:07.005 "is_configured": true, 00:12:07.005 "data_offset": 2048, 00:12:07.005 "data_size": 63488 00:12:07.005 }, 00:12:07.005 { 00:12:07.005 "name": "pt3", 00:12:07.005 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:07.005 "is_configured": true, 00:12:07.005 "data_offset": 2048, 00:12:07.005 "data_size": 63488 00:12:07.005 }, 00:12:07.005 { 00:12:07.005 "name": "pt4", 00:12:07.005 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:07.005 "is_configured": true, 00:12:07.005 "data_offset": 2048, 00:12:07.005 "data_size": 63488 00:12:07.005 } 00:12:07.005 ] 00:12:07.005 } 00:12:07.005 } 00:12:07.005 }' 00:12:07.005 16:00:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:12:07.005 pt2 00:12:07.005 pt3 00:12:07.005 pt4' 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:12:07.005 [2024-11-20 16:00:05.202862] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=e1537747-274c-40d6-8bc8-e51190446046 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z e1537747-274c-40d6-8bc8-e51190446046 ']' 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.005 [2024-11-20 16:00:05.230513] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:07.005 [2024-11-20 16:00:05.230621] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:07.005 [2024-11-20 16:00:05.230718] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:07.005 [2024-11-20 16:00:05.230811] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:07.005 [2024-11-20 16:00:05.230826] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.005 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.266 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:12:07.266 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:12:07.266 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:12:07.266 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:12:07.266 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.267 [2024-11-20 16:00:05.342563] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:12:07.267 [2024-11-20 16:00:05.344459] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:12:07.267 [2024-11-20 16:00:05.344510] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:12:07.267 [2024-11-20 16:00:05.344546] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:12:07.267 [2024-11-20 16:00:05.344595] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:12:07.267 [2024-11-20 16:00:05.344642] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:12:07.267 [2024-11-20 16:00:05.344661] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:12:07.267 [2024-11-20 16:00:05.344694] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:12:07.267 [2024-11-20 16:00:05.344707] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:07.267 [2024-11-20 16:00:05.344718] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:12:07.267 request: 00:12:07.267 { 00:12:07.267 "name": "raid_bdev1", 00:12:07.267 "raid_level": "raid1", 00:12:07.267 "base_bdevs": [ 00:12:07.267 "malloc1", 00:12:07.267 "malloc2", 00:12:07.267 "malloc3", 00:12:07.267 "malloc4" 00:12:07.267 ], 00:12:07.267 "superblock": false, 00:12:07.267 "method": "bdev_raid_create", 00:12:07.267 "req_id": 1 00:12:07.267 } 00:12:07.267 Got JSON-RPC error response 00:12:07.267 response: 00:12:07.267 { 00:12:07.267 "code": -17, 00:12:07.267 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:12:07.267 } 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.267 [2024-11-20 16:00:05.386573] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:12:07.267 [2024-11-20 16:00:05.386631] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:07.267 [2024-11-20 16:00:05.386651] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:12:07.267 [2024-11-20 16:00:05.386662] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:07.267 [2024-11-20 16:00:05.388921] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:07.267 [2024-11-20 16:00:05.388959] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:12:07.267 [2024-11-20 16:00:05.389036] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:12:07.267 [2024-11-20 16:00:05.389087] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:12:07.267 pt1 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.267 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:07.267 "name": "raid_bdev1", 00:12:07.267 "uuid": "e1537747-274c-40d6-8bc8-e51190446046", 00:12:07.267 "strip_size_kb": 0, 00:12:07.267 "state": "configuring", 00:12:07.267 "raid_level": "raid1", 00:12:07.267 "superblock": true, 00:12:07.267 "num_base_bdevs": 4, 00:12:07.267 "num_base_bdevs_discovered": 1, 00:12:07.267 "num_base_bdevs_operational": 4, 00:12:07.267 "base_bdevs_list": [ 00:12:07.267 { 00:12:07.267 "name": "pt1", 00:12:07.267 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:07.267 "is_configured": true, 00:12:07.267 "data_offset": 2048, 00:12:07.267 "data_size": 63488 00:12:07.267 }, 00:12:07.267 { 00:12:07.267 "name": null, 00:12:07.267 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:07.267 "is_configured": false, 00:12:07.267 "data_offset": 2048, 00:12:07.267 "data_size": 63488 00:12:07.267 }, 00:12:07.267 { 00:12:07.267 "name": null, 00:12:07.267 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:07.267 "is_configured": false, 00:12:07.267 "data_offset": 2048, 00:12:07.267 "data_size": 63488 00:12:07.267 }, 00:12:07.267 { 00:12:07.267 "name": null, 00:12:07.267 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:07.267 "is_configured": false, 00:12:07.267 "data_offset": 2048, 00:12:07.267 "data_size": 63488 00:12:07.268 } 00:12:07.268 ] 00:12:07.268 }' 00:12:07.268 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:07.268 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.528 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:12:07.528 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:12:07.528 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.528 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.528 [2024-11-20 16:00:05.718653] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:12:07.528 [2024-11-20 16:00:05.718724] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:07.528 [2024-11-20 16:00:05.718745] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:12:07.528 [2024-11-20 16:00:05.718756] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:07.528 [2024-11-20 16:00:05.719155] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:07.528 [2024-11-20 16:00:05.719172] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:12:07.528 [2024-11-20 16:00:05.719244] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:12:07.528 [2024-11-20 16:00:05.719267] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:07.528 pt2 00:12:07.528 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.529 [2024-11-20 16:00:05.726644] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:07.529 "name": "raid_bdev1", 00:12:07.529 "uuid": "e1537747-274c-40d6-8bc8-e51190446046", 00:12:07.529 "strip_size_kb": 0, 00:12:07.529 "state": "configuring", 00:12:07.529 "raid_level": "raid1", 00:12:07.529 "superblock": true, 00:12:07.529 "num_base_bdevs": 4, 00:12:07.529 "num_base_bdevs_discovered": 1, 00:12:07.529 "num_base_bdevs_operational": 4, 00:12:07.529 "base_bdevs_list": [ 00:12:07.529 { 00:12:07.529 "name": "pt1", 00:12:07.529 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:07.529 "is_configured": true, 00:12:07.529 "data_offset": 2048, 00:12:07.529 "data_size": 63488 00:12:07.529 }, 00:12:07.529 { 00:12:07.529 "name": null, 00:12:07.529 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:07.529 "is_configured": false, 00:12:07.529 "data_offset": 0, 00:12:07.529 "data_size": 63488 00:12:07.529 }, 00:12:07.529 { 00:12:07.529 "name": null, 00:12:07.529 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:07.529 "is_configured": false, 00:12:07.529 "data_offset": 2048, 00:12:07.529 "data_size": 63488 00:12:07.529 }, 00:12:07.529 { 00:12:07.529 "name": null, 00:12:07.529 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:07.529 "is_configured": false, 00:12:07.529 "data_offset": 2048, 00:12:07.529 "data_size": 63488 00:12:07.529 } 00:12:07.529 ] 00:12:07.529 }' 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:07.529 16:00:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.789 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:12:07.789 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:12:07.789 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:12:07.789 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.789 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.789 [2024-11-20 16:00:06.034730] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:12:07.789 [2024-11-20 16:00:06.034789] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:07.789 [2024-11-20 16:00:06.034807] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:12:07.789 [2024-11-20 16:00:06.034816] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:07.789 [2024-11-20 16:00:06.035226] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:07.789 [2024-11-20 16:00:06.035245] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:12:07.789 [2024-11-20 16:00:06.035317] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:12:07.789 [2024-11-20 16:00:06.035341] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:08.051 pt2 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.051 [2024-11-20 16:00:06.042724] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:12:08.051 [2024-11-20 16:00:06.042767] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:08.051 [2024-11-20 16:00:06.042784] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:12:08.051 [2024-11-20 16:00:06.042792] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:08.051 [2024-11-20 16:00:06.043157] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:08.051 [2024-11-20 16:00:06.043174] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:12:08.051 [2024-11-20 16:00:06.043236] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:12:08.051 [2024-11-20 16:00:06.043254] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:12:08.051 pt3 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.051 [2024-11-20 16:00:06.050700] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:12:08.051 [2024-11-20 16:00:06.050735] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:08.051 [2024-11-20 16:00:06.050750] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:12:08.051 [2024-11-20 16:00:06.050758] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:08.051 [2024-11-20 16:00:06.051114] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:08.051 [2024-11-20 16:00:06.051127] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:12:08.051 [2024-11-20 16:00:06.051184] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:12:08.051 [2024-11-20 16:00:06.051203] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:12:08.051 [2024-11-20 16:00:06.051335] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:12:08.051 [2024-11-20 16:00:06.051344] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:08.051 [2024-11-20 16:00:06.051572] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:12:08.051 [2024-11-20 16:00:06.051724] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:12:08.051 [2024-11-20 16:00:06.051735] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:12:08.051 [2024-11-20 16:00:06.051857] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:08.051 pt4 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:08.051 "name": "raid_bdev1", 00:12:08.051 "uuid": "e1537747-274c-40d6-8bc8-e51190446046", 00:12:08.051 "strip_size_kb": 0, 00:12:08.051 "state": "online", 00:12:08.051 "raid_level": "raid1", 00:12:08.051 "superblock": true, 00:12:08.051 "num_base_bdevs": 4, 00:12:08.051 "num_base_bdevs_discovered": 4, 00:12:08.051 "num_base_bdevs_operational": 4, 00:12:08.051 "base_bdevs_list": [ 00:12:08.051 { 00:12:08.051 "name": "pt1", 00:12:08.051 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:08.051 "is_configured": true, 00:12:08.051 "data_offset": 2048, 00:12:08.051 "data_size": 63488 00:12:08.051 }, 00:12:08.051 { 00:12:08.051 "name": "pt2", 00:12:08.051 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:08.051 "is_configured": true, 00:12:08.051 "data_offset": 2048, 00:12:08.051 "data_size": 63488 00:12:08.051 }, 00:12:08.051 { 00:12:08.051 "name": "pt3", 00:12:08.051 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:08.051 "is_configured": true, 00:12:08.051 "data_offset": 2048, 00:12:08.051 "data_size": 63488 00:12:08.051 }, 00:12:08.051 { 00:12:08.051 "name": "pt4", 00:12:08.051 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:08.051 "is_configured": true, 00:12:08.051 "data_offset": 2048, 00:12:08.051 "data_size": 63488 00:12:08.051 } 00:12:08.051 ] 00:12:08.051 }' 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:08.051 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.314 [2024-11-20 16:00:06.387284] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:12:08.314 "name": "raid_bdev1", 00:12:08.314 "aliases": [ 00:12:08.314 "e1537747-274c-40d6-8bc8-e51190446046" 00:12:08.314 ], 00:12:08.314 "product_name": "Raid Volume", 00:12:08.314 "block_size": 512, 00:12:08.314 "num_blocks": 63488, 00:12:08.314 "uuid": "e1537747-274c-40d6-8bc8-e51190446046", 00:12:08.314 "assigned_rate_limits": { 00:12:08.314 "rw_ios_per_sec": 0, 00:12:08.314 "rw_mbytes_per_sec": 0, 00:12:08.314 "r_mbytes_per_sec": 0, 00:12:08.314 "w_mbytes_per_sec": 0 00:12:08.314 }, 00:12:08.314 "claimed": false, 00:12:08.314 "zoned": false, 00:12:08.314 "supported_io_types": { 00:12:08.314 "read": true, 00:12:08.314 "write": true, 00:12:08.314 "unmap": false, 00:12:08.314 "flush": false, 00:12:08.314 "reset": true, 00:12:08.314 "nvme_admin": false, 00:12:08.314 "nvme_io": false, 00:12:08.314 "nvme_io_md": false, 00:12:08.314 "write_zeroes": true, 00:12:08.314 "zcopy": false, 00:12:08.314 "get_zone_info": false, 00:12:08.314 "zone_management": false, 00:12:08.314 "zone_append": false, 00:12:08.314 "compare": false, 00:12:08.314 "compare_and_write": false, 00:12:08.314 "abort": false, 00:12:08.314 "seek_hole": false, 00:12:08.314 "seek_data": false, 00:12:08.314 "copy": false, 00:12:08.314 "nvme_iov_md": false 00:12:08.314 }, 00:12:08.314 "memory_domains": [ 00:12:08.314 { 00:12:08.314 "dma_device_id": "system", 00:12:08.314 "dma_device_type": 1 00:12:08.314 }, 00:12:08.314 { 00:12:08.314 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:08.314 "dma_device_type": 2 00:12:08.314 }, 00:12:08.314 { 00:12:08.314 "dma_device_id": "system", 00:12:08.314 "dma_device_type": 1 00:12:08.314 }, 00:12:08.314 { 00:12:08.314 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:08.314 "dma_device_type": 2 00:12:08.314 }, 00:12:08.314 { 00:12:08.314 "dma_device_id": "system", 00:12:08.314 "dma_device_type": 1 00:12:08.314 }, 00:12:08.314 { 00:12:08.314 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:08.314 "dma_device_type": 2 00:12:08.314 }, 00:12:08.314 { 00:12:08.314 "dma_device_id": "system", 00:12:08.314 "dma_device_type": 1 00:12:08.314 }, 00:12:08.314 { 00:12:08.314 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:08.314 "dma_device_type": 2 00:12:08.314 } 00:12:08.314 ], 00:12:08.314 "driver_specific": { 00:12:08.314 "raid": { 00:12:08.314 "uuid": "e1537747-274c-40d6-8bc8-e51190446046", 00:12:08.314 "strip_size_kb": 0, 00:12:08.314 "state": "online", 00:12:08.314 "raid_level": "raid1", 00:12:08.314 "superblock": true, 00:12:08.314 "num_base_bdevs": 4, 00:12:08.314 "num_base_bdevs_discovered": 4, 00:12:08.314 "num_base_bdevs_operational": 4, 00:12:08.314 "base_bdevs_list": [ 00:12:08.314 { 00:12:08.314 "name": "pt1", 00:12:08.314 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:08.314 "is_configured": true, 00:12:08.314 "data_offset": 2048, 00:12:08.314 "data_size": 63488 00:12:08.314 }, 00:12:08.314 { 00:12:08.314 "name": "pt2", 00:12:08.314 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:08.314 "is_configured": true, 00:12:08.314 "data_offset": 2048, 00:12:08.314 "data_size": 63488 00:12:08.314 }, 00:12:08.314 { 00:12:08.314 "name": "pt3", 00:12:08.314 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:08.314 "is_configured": true, 00:12:08.314 "data_offset": 2048, 00:12:08.314 "data_size": 63488 00:12:08.314 }, 00:12:08.314 { 00:12:08.314 "name": "pt4", 00:12:08.314 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:08.314 "is_configured": true, 00:12:08.314 "data_offset": 2048, 00:12:08.314 "data_size": 63488 00:12:08.314 } 00:12:08.314 ] 00:12:08.314 } 00:12:08.314 } 00:12:08.314 }' 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:12:08.314 pt2 00:12:08.314 pt3 00:12:08.314 pt4' 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.314 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:08.315 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.315 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:08.315 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:08.315 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:08.315 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:08.315 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.576 [2024-11-20 16:00:06.623354] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' e1537747-274c-40d6-8bc8-e51190446046 '!=' e1537747-274c-40d6-8bc8-e51190446046 ']' 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.576 [2024-11-20 16:00:06.651101] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.576 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:08.577 "name": "raid_bdev1", 00:12:08.577 "uuid": "e1537747-274c-40d6-8bc8-e51190446046", 00:12:08.577 "strip_size_kb": 0, 00:12:08.577 "state": "online", 00:12:08.577 "raid_level": "raid1", 00:12:08.577 "superblock": true, 00:12:08.577 "num_base_bdevs": 4, 00:12:08.577 "num_base_bdevs_discovered": 3, 00:12:08.577 "num_base_bdevs_operational": 3, 00:12:08.577 "base_bdevs_list": [ 00:12:08.577 { 00:12:08.577 "name": null, 00:12:08.577 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:08.577 "is_configured": false, 00:12:08.577 "data_offset": 0, 00:12:08.577 "data_size": 63488 00:12:08.577 }, 00:12:08.577 { 00:12:08.577 "name": "pt2", 00:12:08.577 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:08.577 "is_configured": true, 00:12:08.577 "data_offset": 2048, 00:12:08.577 "data_size": 63488 00:12:08.577 }, 00:12:08.577 { 00:12:08.577 "name": "pt3", 00:12:08.577 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:08.577 "is_configured": true, 00:12:08.577 "data_offset": 2048, 00:12:08.577 "data_size": 63488 00:12:08.577 }, 00:12:08.577 { 00:12:08.577 "name": "pt4", 00:12:08.577 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:08.577 "is_configured": true, 00:12:08.577 "data_offset": 2048, 00:12:08.577 "data_size": 63488 00:12:08.577 } 00:12:08.577 ] 00:12:08.577 }' 00:12:08.577 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:08.577 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.838 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:08.838 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.838 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.838 [2024-11-20 16:00:06.975211] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:08.838 [2024-11-20 16:00:06.975240] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:08.838 [2024-11-20 16:00:06.975307] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:08.839 [2024-11-20 16:00:06.975385] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:08.839 [2024-11-20 16:00:06.975395] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:12:08.839 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.839 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:12:08.839 16:00:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:08.839 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.839 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.839 16:00:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.839 [2024-11-20 16:00:07.047242] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:12:08.839 [2024-11-20 16:00:07.047290] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:08.839 [2024-11-20 16:00:07.047308] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:12:08.839 [2024-11-20 16:00:07.047317] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:08.839 [2024-11-20 16:00:07.049503] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:08.839 [2024-11-20 16:00:07.049624] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:12:08.839 [2024-11-20 16:00:07.049723] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:12:08.839 [2024-11-20 16:00:07.049767] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:08.839 pt2 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:08.839 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.100 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:09.100 "name": "raid_bdev1", 00:12:09.100 "uuid": "e1537747-274c-40d6-8bc8-e51190446046", 00:12:09.100 "strip_size_kb": 0, 00:12:09.100 "state": "configuring", 00:12:09.100 "raid_level": "raid1", 00:12:09.100 "superblock": true, 00:12:09.100 "num_base_bdevs": 4, 00:12:09.100 "num_base_bdevs_discovered": 1, 00:12:09.100 "num_base_bdevs_operational": 3, 00:12:09.100 "base_bdevs_list": [ 00:12:09.100 { 00:12:09.100 "name": null, 00:12:09.100 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:09.100 "is_configured": false, 00:12:09.100 "data_offset": 2048, 00:12:09.100 "data_size": 63488 00:12:09.100 }, 00:12:09.100 { 00:12:09.100 "name": "pt2", 00:12:09.100 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:09.100 "is_configured": true, 00:12:09.100 "data_offset": 2048, 00:12:09.100 "data_size": 63488 00:12:09.100 }, 00:12:09.100 { 00:12:09.100 "name": null, 00:12:09.100 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:09.100 "is_configured": false, 00:12:09.100 "data_offset": 2048, 00:12:09.100 "data_size": 63488 00:12:09.100 }, 00:12:09.100 { 00:12:09.100 "name": null, 00:12:09.100 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:09.100 "is_configured": false, 00:12:09.100 "data_offset": 2048, 00:12:09.100 "data_size": 63488 00:12:09.100 } 00:12:09.100 ] 00:12:09.100 }' 00:12:09.100 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:09.100 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.361 [2024-11-20 16:00:07.363443] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:12:09.361 [2024-11-20 16:00:07.363499] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:09.361 [2024-11-20 16:00:07.363518] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:12:09.361 [2024-11-20 16:00:07.363527] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:09.361 [2024-11-20 16:00:07.363957] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:09.361 [2024-11-20 16:00:07.363976] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:12:09.361 [2024-11-20 16:00:07.364052] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:12:09.361 [2024-11-20 16:00:07.364071] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:12:09.361 pt3 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.361 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:09.361 "name": "raid_bdev1", 00:12:09.361 "uuid": "e1537747-274c-40d6-8bc8-e51190446046", 00:12:09.361 "strip_size_kb": 0, 00:12:09.361 "state": "configuring", 00:12:09.361 "raid_level": "raid1", 00:12:09.361 "superblock": true, 00:12:09.361 "num_base_bdevs": 4, 00:12:09.362 "num_base_bdevs_discovered": 2, 00:12:09.362 "num_base_bdevs_operational": 3, 00:12:09.362 "base_bdevs_list": [ 00:12:09.362 { 00:12:09.362 "name": null, 00:12:09.362 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:09.362 "is_configured": false, 00:12:09.362 "data_offset": 2048, 00:12:09.362 "data_size": 63488 00:12:09.362 }, 00:12:09.362 { 00:12:09.362 "name": "pt2", 00:12:09.362 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:09.362 "is_configured": true, 00:12:09.362 "data_offset": 2048, 00:12:09.362 "data_size": 63488 00:12:09.362 }, 00:12:09.362 { 00:12:09.362 "name": "pt3", 00:12:09.362 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:09.362 "is_configured": true, 00:12:09.362 "data_offset": 2048, 00:12:09.362 "data_size": 63488 00:12:09.362 }, 00:12:09.362 { 00:12:09.362 "name": null, 00:12:09.362 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:09.362 "is_configured": false, 00:12:09.362 "data_offset": 2048, 00:12:09.362 "data_size": 63488 00:12:09.362 } 00:12:09.362 ] 00:12:09.362 }' 00:12:09.362 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:09.362 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.623 [2024-11-20 16:00:07.695633] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:12:09.623 [2024-11-20 16:00:07.695809] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:09.623 [2024-11-20 16:00:07.695841] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:12:09.623 [2024-11-20 16:00:07.695851] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:09.623 [2024-11-20 16:00:07.696263] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:09.623 [2024-11-20 16:00:07.696276] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:12:09.623 [2024-11-20 16:00:07.696351] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:12:09.623 [2024-11-20 16:00:07.696372] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:12:09.623 [2024-11-20 16:00:07.696499] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:12:09.623 [2024-11-20 16:00:07.696508] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:09.623 [2024-11-20 16:00:07.696759] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:12:09.623 [2024-11-20 16:00:07.696897] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:12:09.623 [2024-11-20 16:00:07.696907] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:12:09.623 [2024-11-20 16:00:07.697033] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:09.623 pt4 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.623 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:09.623 "name": "raid_bdev1", 00:12:09.623 "uuid": "e1537747-274c-40d6-8bc8-e51190446046", 00:12:09.623 "strip_size_kb": 0, 00:12:09.623 "state": "online", 00:12:09.623 "raid_level": "raid1", 00:12:09.623 "superblock": true, 00:12:09.623 "num_base_bdevs": 4, 00:12:09.623 "num_base_bdevs_discovered": 3, 00:12:09.623 "num_base_bdevs_operational": 3, 00:12:09.623 "base_bdevs_list": [ 00:12:09.623 { 00:12:09.623 "name": null, 00:12:09.623 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:09.623 "is_configured": false, 00:12:09.623 "data_offset": 2048, 00:12:09.623 "data_size": 63488 00:12:09.623 }, 00:12:09.623 { 00:12:09.623 "name": "pt2", 00:12:09.623 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:09.623 "is_configured": true, 00:12:09.623 "data_offset": 2048, 00:12:09.623 "data_size": 63488 00:12:09.623 }, 00:12:09.623 { 00:12:09.623 "name": "pt3", 00:12:09.623 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:09.623 "is_configured": true, 00:12:09.623 "data_offset": 2048, 00:12:09.623 "data_size": 63488 00:12:09.623 }, 00:12:09.623 { 00:12:09.623 "name": "pt4", 00:12:09.623 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:09.623 "is_configured": true, 00:12:09.623 "data_offset": 2048, 00:12:09.623 "data_size": 63488 00:12:09.624 } 00:12:09.624 ] 00:12:09.624 }' 00:12:09.624 16:00:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:09.624 16:00:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.934 [2024-11-20 16:00:08.015768] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:09.934 [2024-11-20 16:00:08.015793] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:09.934 [2024-11-20 16:00:08.015860] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:09.934 [2024-11-20 16:00:08.015932] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:09.934 [2024-11-20 16:00:08.015944] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.934 [2024-11-20 16:00:08.067838] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:12:09.934 [2024-11-20 16:00:08.067889] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:09.934 [2024-11-20 16:00:08.067904] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:12:09.934 [2024-11-20 16:00:08.067916] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:09.934 [2024-11-20 16:00:08.070211] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:09.934 [2024-11-20 16:00:08.070245] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:12:09.934 [2024-11-20 16:00:08.070317] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:12:09.934 [2024-11-20 16:00:08.070358] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:12:09.934 [2024-11-20 16:00:08.070476] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:12:09.934 [2024-11-20 16:00:08.070489] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:09.934 [2024-11-20 16:00:08.070504] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:12:09.934 [2024-11-20 16:00:08.070557] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:09.934 [2024-11-20 16:00:08.070652] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:12:09.934 pt1 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.934 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:09.934 "name": "raid_bdev1", 00:12:09.934 "uuid": "e1537747-274c-40d6-8bc8-e51190446046", 00:12:09.934 "strip_size_kb": 0, 00:12:09.934 "state": "configuring", 00:12:09.934 "raid_level": "raid1", 00:12:09.934 "superblock": true, 00:12:09.934 "num_base_bdevs": 4, 00:12:09.934 "num_base_bdevs_discovered": 2, 00:12:09.934 "num_base_bdevs_operational": 3, 00:12:09.934 "base_bdevs_list": [ 00:12:09.934 { 00:12:09.934 "name": null, 00:12:09.934 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:09.934 "is_configured": false, 00:12:09.934 "data_offset": 2048, 00:12:09.934 "data_size": 63488 00:12:09.934 }, 00:12:09.934 { 00:12:09.935 "name": "pt2", 00:12:09.935 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:09.935 "is_configured": true, 00:12:09.935 "data_offset": 2048, 00:12:09.935 "data_size": 63488 00:12:09.935 }, 00:12:09.935 { 00:12:09.935 "name": "pt3", 00:12:09.935 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:09.935 "is_configured": true, 00:12:09.935 "data_offset": 2048, 00:12:09.935 "data_size": 63488 00:12:09.935 }, 00:12:09.935 { 00:12:09.935 "name": null, 00:12:09.935 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:09.935 "is_configured": false, 00:12:09.935 "data_offset": 2048, 00:12:09.935 "data_size": 63488 00:12:09.935 } 00:12:09.935 ] 00:12:09.935 }' 00:12:09.935 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:09.935 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.243 [2024-11-20 16:00:08.440014] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:12:10.243 [2024-11-20 16:00:08.440067] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:10.243 [2024-11-20 16:00:08.440089] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:12:10.243 [2024-11-20 16:00:08.440098] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:10.243 [2024-11-20 16:00:08.440509] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:10.243 [2024-11-20 16:00:08.440530] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:12:10.243 [2024-11-20 16:00:08.440608] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:12:10.243 [2024-11-20 16:00:08.440629] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:12:10.243 [2024-11-20 16:00:08.440762] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:12:10.243 [2024-11-20 16:00:08.440776] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:10.243 [2024-11-20 16:00:08.441024] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:12:10.243 [2024-11-20 16:00:08.441160] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:12:10.243 [2024-11-20 16:00:08.441171] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:12:10.243 [2024-11-20 16:00:08.441296] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:10.243 pt4 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:10.243 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:10.243 "name": "raid_bdev1", 00:12:10.243 "uuid": "e1537747-274c-40d6-8bc8-e51190446046", 00:12:10.243 "strip_size_kb": 0, 00:12:10.243 "state": "online", 00:12:10.243 "raid_level": "raid1", 00:12:10.243 "superblock": true, 00:12:10.243 "num_base_bdevs": 4, 00:12:10.243 "num_base_bdevs_discovered": 3, 00:12:10.243 "num_base_bdevs_operational": 3, 00:12:10.243 "base_bdevs_list": [ 00:12:10.243 { 00:12:10.243 "name": null, 00:12:10.243 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:10.243 "is_configured": false, 00:12:10.243 "data_offset": 2048, 00:12:10.243 "data_size": 63488 00:12:10.243 }, 00:12:10.243 { 00:12:10.243 "name": "pt2", 00:12:10.243 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:10.243 "is_configured": true, 00:12:10.243 "data_offset": 2048, 00:12:10.243 "data_size": 63488 00:12:10.243 }, 00:12:10.243 { 00:12:10.243 "name": "pt3", 00:12:10.243 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:10.243 "is_configured": true, 00:12:10.243 "data_offset": 2048, 00:12:10.243 "data_size": 63488 00:12:10.243 }, 00:12:10.243 { 00:12:10.243 "name": "pt4", 00:12:10.243 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:10.243 "is_configured": true, 00:12:10.243 "data_offset": 2048, 00:12:10.243 "data_size": 63488 00:12:10.244 } 00:12:10.244 ] 00:12:10.244 }' 00:12:10.244 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:10.244 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.812 [2024-11-20 16:00:08.796508] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' e1537747-274c-40d6-8bc8-e51190446046 '!=' e1537747-274c-40d6-8bc8-e51190446046 ']' 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 72431 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 72431 ']' 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 72431 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72431 00:12:10.812 killing process with pid 72431 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72431' 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 72431 00:12:10.812 [2024-11-20 16:00:08.847912] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:10.812 [2024-11-20 16:00:08.847994] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:10.812 [2024-11-20 16:00:08.848067] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:10.812 16:00:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 72431 00:12:10.812 [2024-11-20 16:00:08.848079] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:12:11.077 [2024-11-20 16:00:09.091170] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:11.649 ************************************ 00:12:11.649 END TEST raid_superblock_test 00:12:11.649 ************************************ 00:12:11.649 16:00:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:12:11.649 00:12:11.649 real 0m6.281s 00:12:11.649 user 0m9.886s 00:12:11.649 sys 0m1.020s 00:12:11.649 16:00:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:11.649 16:00:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.649 16:00:09 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 4 read 00:12:11.649 16:00:09 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:12:11.649 16:00:09 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:11.649 16:00:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:11.649 ************************************ 00:12:11.649 START TEST raid_read_error_test 00:12:11.649 ************************************ 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 4 read 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.2EHnUabfkw 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=72901 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 72901 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 72901 ']' 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:11.649 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:12:11.649 16:00:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:11.911 16:00:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:11.911 16:00:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.911 [2024-11-20 16:00:09.961824] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:12:11.911 [2024-11-20 16:00:09.962001] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72901 ] 00:12:11.911 [2024-11-20 16:00:10.128754] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:12.173 [2024-11-20 16:00:10.233042] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:12.173 [2024-11-20 16:00:10.371613] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:12.173 [2024-11-20 16:00:10.371681] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:12.744 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:12.744 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:12:12.744 16:00:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:12:12.744 16:00:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:12.744 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.744 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:12.744 BaseBdev1_malloc 00:12:12.744 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.744 16:00:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:12:12.744 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.744 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:12.744 true 00:12:12.744 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.744 16:00:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:12.745 [2024-11-20 16:00:10.897866] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:12:12.745 [2024-11-20 16:00:10.897922] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:12.745 [2024-11-20 16:00:10.897941] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:12:12.745 [2024-11-20 16:00:10.897952] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:12.745 [2024-11-20 16:00:10.900142] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:12.745 [2024-11-20 16:00:10.900181] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:12.745 BaseBdev1 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:12.745 BaseBdev2_malloc 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:12.745 true 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:12.745 [2024-11-20 16:00:10.954366] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:12:12.745 [2024-11-20 16:00:10.954428] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:12.745 [2024-11-20 16:00:10.954447] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:12:12.745 [2024-11-20 16:00:10.954458] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:12.745 [2024-11-20 16:00:10.956618] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:12.745 [2024-11-20 16:00:10.956655] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:12.745 BaseBdev2 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.745 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.005 BaseBdev3_malloc 00:12:13.005 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.005 16:00:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:12:13.005 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.005 16:00:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.005 true 00:12:13.005 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.005 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:12:13.005 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.005 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.005 [2024-11-20 16:00:11.016830] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:12:13.005 [2024-11-20 16:00:11.016891] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:13.005 [2024-11-20 16:00:11.016911] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:12:13.005 [2024-11-20 16:00:11.016922] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:13.005 [2024-11-20 16:00:11.019119] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:13.005 [2024-11-20 16:00:11.019156] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:13.005 BaseBdev3 00:12:13.005 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.005 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:12:13.005 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:13.005 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.005 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.005 BaseBdev4_malloc 00:12:13.005 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.005 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:12:13.005 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.006 true 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.006 [2024-11-20 16:00:11.065495] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:12:13.006 [2024-11-20 16:00:11.065554] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:13.006 [2024-11-20 16:00:11.065574] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:12:13.006 [2024-11-20 16:00:11.065584] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:13.006 [2024-11-20 16:00:11.067763] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:13.006 [2024-11-20 16:00:11.067803] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:13.006 BaseBdev4 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.006 [2024-11-20 16:00:11.073557] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:13.006 [2024-11-20 16:00:11.075428] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:13.006 [2024-11-20 16:00:11.075510] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:13.006 [2024-11-20 16:00:11.075578] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:13.006 [2024-11-20 16:00:11.075827] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:12:13.006 [2024-11-20 16:00:11.075847] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:13.006 [2024-11-20 16:00:11.076112] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:12:13.006 [2024-11-20 16:00:11.076272] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:12:13.006 [2024-11-20 16:00:11.076288] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:12:13.006 [2024-11-20 16:00:11.076446] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:13.006 "name": "raid_bdev1", 00:12:13.006 "uuid": "ae96e40a-8bc3-4b27-8684-623ee819e734", 00:12:13.006 "strip_size_kb": 0, 00:12:13.006 "state": "online", 00:12:13.006 "raid_level": "raid1", 00:12:13.006 "superblock": true, 00:12:13.006 "num_base_bdevs": 4, 00:12:13.006 "num_base_bdevs_discovered": 4, 00:12:13.006 "num_base_bdevs_operational": 4, 00:12:13.006 "base_bdevs_list": [ 00:12:13.006 { 00:12:13.006 "name": "BaseBdev1", 00:12:13.006 "uuid": "a7df852a-dc25-58ae-8428-da5e57180378", 00:12:13.006 "is_configured": true, 00:12:13.006 "data_offset": 2048, 00:12:13.006 "data_size": 63488 00:12:13.006 }, 00:12:13.006 { 00:12:13.006 "name": "BaseBdev2", 00:12:13.006 "uuid": "751a3757-5ac3-5aa1-8dab-8c12ca29420f", 00:12:13.006 "is_configured": true, 00:12:13.006 "data_offset": 2048, 00:12:13.006 "data_size": 63488 00:12:13.006 }, 00:12:13.006 { 00:12:13.006 "name": "BaseBdev3", 00:12:13.006 "uuid": "5bc488de-4e38-5ff9-8bb5-14eb16dd7e5d", 00:12:13.006 "is_configured": true, 00:12:13.006 "data_offset": 2048, 00:12:13.006 "data_size": 63488 00:12:13.006 }, 00:12:13.006 { 00:12:13.006 "name": "BaseBdev4", 00:12:13.006 "uuid": "fde308fa-e30e-5b59-b09e-79ad32937b30", 00:12:13.006 "is_configured": true, 00:12:13.006 "data_offset": 2048, 00:12:13.006 "data_size": 63488 00:12:13.006 } 00:12:13.006 ] 00:12:13.006 }' 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:13.006 16:00:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.266 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:12:13.266 16:00:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:13.266 [2024-11-20 16:00:11.494657] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006a40 00:12:14.285 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:12:14.285 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.285 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:14.285 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.285 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:12:14.285 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:12:14.285 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:12:14.285 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:14.286 "name": "raid_bdev1", 00:12:14.286 "uuid": "ae96e40a-8bc3-4b27-8684-623ee819e734", 00:12:14.286 "strip_size_kb": 0, 00:12:14.286 "state": "online", 00:12:14.286 "raid_level": "raid1", 00:12:14.286 "superblock": true, 00:12:14.286 "num_base_bdevs": 4, 00:12:14.286 "num_base_bdevs_discovered": 4, 00:12:14.286 "num_base_bdevs_operational": 4, 00:12:14.286 "base_bdevs_list": [ 00:12:14.286 { 00:12:14.286 "name": "BaseBdev1", 00:12:14.286 "uuid": "a7df852a-dc25-58ae-8428-da5e57180378", 00:12:14.286 "is_configured": true, 00:12:14.286 "data_offset": 2048, 00:12:14.286 "data_size": 63488 00:12:14.286 }, 00:12:14.286 { 00:12:14.286 "name": "BaseBdev2", 00:12:14.286 "uuid": "751a3757-5ac3-5aa1-8dab-8c12ca29420f", 00:12:14.286 "is_configured": true, 00:12:14.286 "data_offset": 2048, 00:12:14.286 "data_size": 63488 00:12:14.286 }, 00:12:14.286 { 00:12:14.286 "name": "BaseBdev3", 00:12:14.286 "uuid": "5bc488de-4e38-5ff9-8bb5-14eb16dd7e5d", 00:12:14.286 "is_configured": true, 00:12:14.286 "data_offset": 2048, 00:12:14.286 "data_size": 63488 00:12:14.286 }, 00:12:14.286 { 00:12:14.286 "name": "BaseBdev4", 00:12:14.286 "uuid": "fde308fa-e30e-5b59-b09e-79ad32937b30", 00:12:14.286 "is_configured": true, 00:12:14.286 "data_offset": 2048, 00:12:14.286 "data_size": 63488 00:12:14.286 } 00:12:14.286 ] 00:12:14.286 }' 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:14.286 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:14.546 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:14.546 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:14.546 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:14.546 [2024-11-20 16:00:12.770817] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:14.546 [2024-11-20 16:00:12.770849] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:14.546 [2024-11-20 16:00:12.774100] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:14.546 [2024-11-20 16:00:12.774167] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:14.546 [2024-11-20 16:00:12.774317] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:14.546 [2024-11-20 16:00:12.774332] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:12:14.546 { 00:12:14.546 "results": [ 00:12:14.546 { 00:12:14.546 "job": "raid_bdev1", 00:12:14.546 "core_mask": "0x1", 00:12:14.546 "workload": "randrw", 00:12:14.546 "percentage": 50, 00:12:14.546 "status": "finished", 00:12:14.546 "queue_depth": 1, 00:12:14.546 "io_size": 131072, 00:12:14.546 "runtime": 1.274359, 00:12:14.546 "iops": 11080.080259958144, 00:12:14.546 "mibps": 1385.010032494768, 00:12:14.546 "io_failed": 0, 00:12:14.546 "io_timeout": 0, 00:12:14.546 "avg_latency_us": 86.99029505338854, 00:12:14.546 "min_latency_us": 29.932307692307692, 00:12:14.546 "max_latency_us": 1777.033846153846 00:12:14.546 } 00:12:14.546 ], 00:12:14.546 "core_count": 1 00:12:14.546 } 00:12:14.546 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:14.546 16:00:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 72901 00:12:14.546 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 72901 ']' 00:12:14.546 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 72901 00:12:14.546 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:12:14.546 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:14.546 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72901 00:12:14.807 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:14.807 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:14.807 killing process with pid 72901 00:12:14.807 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72901' 00:12:14.807 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 72901 00:12:14.807 [2024-11-20 16:00:12.804364] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:14.807 16:00:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 72901 00:12:14.807 [2024-11-20 16:00:13.009416] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:15.750 16:00:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.2EHnUabfkw 00:12:15.750 16:00:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:12:15.750 16:00:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:12:15.750 16:00:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:12:15.750 16:00:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:12:15.750 16:00:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:12:15.750 16:00:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:12:15.750 16:00:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:12:15.750 00:12:15.750 real 0m3.897s 00:12:15.750 user 0m4.619s 00:12:15.750 sys 0m0.433s 00:12:15.750 16:00:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:15.750 16:00:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:15.750 ************************************ 00:12:15.750 END TEST raid_read_error_test 00:12:15.750 ************************************ 00:12:15.750 16:00:13 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 4 write 00:12:15.750 16:00:13 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:12:15.750 16:00:13 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:15.750 16:00:13 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:15.750 ************************************ 00:12:15.750 START TEST raid_write_error_test 00:12:15.750 ************************************ 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 4 write 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.QmYIWCbkGV 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73041 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73041 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 73041 ']' 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:15.750 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:15.750 16:00:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:15.750 [2024-11-20 16:00:13.926172] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:12:15.750 [2024-11-20 16:00:13.926301] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73041 ] 00:12:16.009 [2024-11-20 16:00:14.084433] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:16.009 [2024-11-20 16:00:14.193311] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:16.268 [2024-11-20 16:00:14.331493] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:16.268 [2024-11-20 16:00:14.331552] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:16.839 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:16.839 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:12:16.839 16:00:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:12:16.839 16:00:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:16.839 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.839 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:16.839 BaseBdev1_malloc 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:16.840 true 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:16.840 [2024-11-20 16:00:14.918528] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:12:16.840 [2024-11-20 16:00:14.918587] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:16.840 [2024-11-20 16:00:14.918611] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:12:16.840 [2024-11-20 16:00:14.918624] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:16.840 [2024-11-20 16:00:14.920821] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:16.840 [2024-11-20 16:00:14.920860] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:16.840 BaseBdev1 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:16.840 BaseBdev2_malloc 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:16.840 true 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:16.840 [2024-11-20 16:00:14.976469] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:12:16.840 [2024-11-20 16:00:14.976530] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:16.840 [2024-11-20 16:00:14.976548] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:12:16.840 [2024-11-20 16:00:14.976558] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:16.840 [2024-11-20 16:00:14.978832] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:16.840 [2024-11-20 16:00:14.978871] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:16.840 BaseBdev2 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.840 16:00:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:16.840 BaseBdev3_malloc 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:16.840 true 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:16.840 [2024-11-20 16:00:15.034678] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:12:16.840 [2024-11-20 16:00:15.034736] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:16.840 [2024-11-20 16:00:15.034757] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:12:16.840 [2024-11-20 16:00:15.034769] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:16.840 [2024-11-20 16:00:15.037245] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:16.840 [2024-11-20 16:00:15.037296] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:16.840 BaseBdev3 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:16.840 BaseBdev4_malloc 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:16.840 true 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:16.840 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:16.840 [2024-11-20 16:00:15.083416] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:12:16.840 [2024-11-20 16:00:15.083469] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:16.840 [2024-11-20 16:00:15.083499] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:12:16.840 [2024-11-20 16:00:15.083509] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:16.840 [2024-11-20 16:00:15.085657] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:16.840 [2024-11-20 16:00:15.085709] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:17.100 BaseBdev4 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:17.100 [2024-11-20 16:00:15.095500] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:17.100 [2024-11-20 16:00:15.097371] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:17.100 [2024-11-20 16:00:15.097455] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:17.100 [2024-11-20 16:00:15.097522] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:17.100 [2024-11-20 16:00:15.097765] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:12:17.100 [2024-11-20 16:00:15.097786] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:17.100 [2024-11-20 16:00:15.098052] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:12:17.100 [2024-11-20 16:00:15.098209] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:12:17.100 [2024-11-20 16:00:15.098218] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:12:17.100 [2024-11-20 16:00:15.098372] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:17.100 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:17.100 "name": "raid_bdev1", 00:12:17.100 "uuid": "1218ceed-0209-45e5-afde-1c0efc2b9a78", 00:12:17.100 "strip_size_kb": 0, 00:12:17.100 "state": "online", 00:12:17.100 "raid_level": "raid1", 00:12:17.100 "superblock": true, 00:12:17.100 "num_base_bdevs": 4, 00:12:17.100 "num_base_bdevs_discovered": 4, 00:12:17.100 "num_base_bdevs_operational": 4, 00:12:17.100 "base_bdevs_list": [ 00:12:17.100 { 00:12:17.101 "name": "BaseBdev1", 00:12:17.101 "uuid": "898f64b1-ad7b-5b19-8f24-ff08fc197bb3", 00:12:17.101 "is_configured": true, 00:12:17.101 "data_offset": 2048, 00:12:17.101 "data_size": 63488 00:12:17.101 }, 00:12:17.101 { 00:12:17.101 "name": "BaseBdev2", 00:12:17.101 "uuid": "d14d25b7-e98d-5d9b-97d1-42689cb53e93", 00:12:17.101 "is_configured": true, 00:12:17.101 "data_offset": 2048, 00:12:17.101 "data_size": 63488 00:12:17.101 }, 00:12:17.101 { 00:12:17.101 "name": "BaseBdev3", 00:12:17.101 "uuid": "d24cb069-da93-5b43-abd0-7bfcf6d5bb62", 00:12:17.101 "is_configured": true, 00:12:17.101 "data_offset": 2048, 00:12:17.101 "data_size": 63488 00:12:17.101 }, 00:12:17.101 { 00:12:17.101 "name": "BaseBdev4", 00:12:17.101 "uuid": "d41b9f9c-2e01-5be2-a46a-81c8029abb33", 00:12:17.101 "is_configured": true, 00:12:17.101 "data_offset": 2048, 00:12:17.101 "data_size": 63488 00:12:17.101 } 00:12:17.101 ] 00:12:17.101 }' 00:12:17.101 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:17.101 16:00:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:17.361 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:12:17.361 16:00:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:17.361 [2024-11-20 16:00:15.552536] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006a40 00:12:18.303 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:12:18.303 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:18.303 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:18.304 [2024-11-20 16:00:16.475885] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:12:18.304 [2024-11-20 16:00:16.475941] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:18.304 [2024-11-20 16:00:16.476165] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006a40 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=3 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:18.304 "name": "raid_bdev1", 00:12:18.304 "uuid": "1218ceed-0209-45e5-afde-1c0efc2b9a78", 00:12:18.304 "strip_size_kb": 0, 00:12:18.304 "state": "online", 00:12:18.304 "raid_level": "raid1", 00:12:18.304 "superblock": true, 00:12:18.304 "num_base_bdevs": 4, 00:12:18.304 "num_base_bdevs_discovered": 3, 00:12:18.304 "num_base_bdevs_operational": 3, 00:12:18.304 "base_bdevs_list": [ 00:12:18.304 { 00:12:18.304 "name": null, 00:12:18.304 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:18.304 "is_configured": false, 00:12:18.304 "data_offset": 0, 00:12:18.304 "data_size": 63488 00:12:18.304 }, 00:12:18.304 { 00:12:18.304 "name": "BaseBdev2", 00:12:18.304 "uuid": "d14d25b7-e98d-5d9b-97d1-42689cb53e93", 00:12:18.304 "is_configured": true, 00:12:18.304 "data_offset": 2048, 00:12:18.304 "data_size": 63488 00:12:18.304 }, 00:12:18.304 { 00:12:18.304 "name": "BaseBdev3", 00:12:18.304 "uuid": "d24cb069-da93-5b43-abd0-7bfcf6d5bb62", 00:12:18.304 "is_configured": true, 00:12:18.304 "data_offset": 2048, 00:12:18.304 "data_size": 63488 00:12:18.304 }, 00:12:18.304 { 00:12:18.304 "name": "BaseBdev4", 00:12:18.304 "uuid": "d41b9f9c-2e01-5be2-a46a-81c8029abb33", 00:12:18.304 "is_configured": true, 00:12:18.304 "data_offset": 2048, 00:12:18.304 "data_size": 63488 00:12:18.304 } 00:12:18.304 ] 00:12:18.304 }' 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:18.304 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:18.877 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:18.877 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:18.877 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:18.877 [2024-11-20 16:00:16.855534] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:18.877 [2024-11-20 16:00:16.855569] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:18.877 [2024-11-20 16:00:16.858763] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:18.877 [2024-11-20 16:00:16.858810] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:18.877 [2024-11-20 16:00:16.858917] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:18.877 [2024-11-20 16:00:16.858930] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:12:18.877 { 00:12:18.877 "results": [ 00:12:18.877 { 00:12:18.877 "job": "raid_bdev1", 00:12:18.877 "core_mask": "0x1", 00:12:18.877 "workload": "randrw", 00:12:18.877 "percentage": 50, 00:12:18.877 "status": "finished", 00:12:18.877 "queue_depth": 1, 00:12:18.877 "io_size": 131072, 00:12:18.877 "runtime": 1.301296, 00:12:18.877 "iops": 11664.525211788863, 00:12:18.877 "mibps": 1458.0656514736079, 00:12:18.877 "io_failed": 0, 00:12:18.877 "io_timeout": 0, 00:12:18.877 "avg_latency_us": 82.47632488204859, 00:12:18.877 "min_latency_us": 30.12923076923077, 00:12:18.877 "max_latency_us": 1777.033846153846 00:12:18.877 } 00:12:18.877 ], 00:12:18.877 "core_count": 1 00:12:18.877 } 00:12:18.877 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:18.877 16:00:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73041 00:12:18.877 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 73041 ']' 00:12:18.877 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 73041 00:12:18.877 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:12:18.877 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:18.877 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73041 00:12:18.877 killing process with pid 73041 00:12:18.877 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:18.877 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:18.877 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73041' 00:12:18.877 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 73041 00:12:18.877 [2024-11-20 16:00:16.888458] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:18.877 16:00:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 73041 00:12:18.877 [2024-11-20 16:00:17.094043] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:19.818 16:00:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:12:19.819 16:00:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.QmYIWCbkGV 00:12:19.819 16:00:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:12:19.819 16:00:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:12:19.819 16:00:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:12:19.819 16:00:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:12:19.819 16:00:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:12:19.819 16:00:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:12:19.819 00:12:19.819 real 0m4.019s 00:12:19.819 user 0m4.846s 00:12:19.819 sys 0m0.447s 00:12:19.819 ************************************ 00:12:19.819 END TEST raid_write_error_test 00:12:19.819 ************************************ 00:12:19.819 16:00:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:19.819 16:00:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.819 16:00:17 bdev_raid -- bdev/bdev_raid.sh@976 -- # '[' true = true ']' 00:12:19.819 16:00:17 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:12:19.819 16:00:17 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 2 false false true 00:12:19.819 16:00:17 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:19.819 16:00:17 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:19.819 16:00:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:19.819 ************************************ 00:12:19.819 START TEST raid_rebuild_test 00:12:19.819 ************************************ 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 false false true 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=73179 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 73179 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 73179 ']' 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:19.819 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:19.819 16:00:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.819 [2024-11-20 16:00:18.009197] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:12:19.819 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:19.819 Zero copy mechanism will not be used. 00:12:19.819 [2024-11-20 16:00:18.009327] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73179 ] 00:12:20.084 [2024-11-20 16:00:18.170157] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:20.084 [2024-11-20 16:00:18.273203] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:20.350 [2024-11-20 16:00:18.411374] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:20.350 [2024-11-20 16:00:18.411435] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:20.609 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:20.609 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:12:20.610 16:00:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:20.610 16:00:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:20.610 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.610 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.868 BaseBdev1_malloc 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.868 [2024-11-20 16:00:18.895309] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:20.868 [2024-11-20 16:00:18.895376] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:20.868 [2024-11-20 16:00:18.895400] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:20.868 [2024-11-20 16:00:18.895413] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:20.868 [2024-11-20 16:00:18.897608] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:20.868 [2024-11-20 16:00:18.897646] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:20.868 BaseBdev1 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.868 BaseBdev2_malloc 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.868 [2024-11-20 16:00:18.939579] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:20.868 [2024-11-20 16:00:18.939647] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:20.868 [2024-11-20 16:00:18.939684] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:20.868 [2024-11-20 16:00:18.939696] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:20.868 [2024-11-20 16:00:18.941898] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:20.868 [2024-11-20 16:00:18.941934] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:20.868 BaseBdev2 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.868 spare_malloc 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.868 spare_delay 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.868 16:00:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.868 [2024-11-20 16:00:19.005056] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:20.868 [2024-11-20 16:00:19.005125] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:20.868 [2024-11-20 16:00:19.005146] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:12:20.868 [2024-11-20 16:00:19.005157] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:20.868 [2024-11-20 16:00:19.007389] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:20.868 [2024-11-20 16:00:19.007431] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:20.868 spare 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.868 [2024-11-20 16:00:19.013102] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:20.868 [2024-11-20 16:00:19.014972] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:20.868 [2024-11-20 16:00:19.015073] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:12:20.868 [2024-11-20 16:00:19.015087] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:12:20.868 [2024-11-20 16:00:19.015356] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:12:20.868 [2024-11-20 16:00:19.015517] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:12:20.868 [2024-11-20 16:00:19.015535] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:12:20.868 [2024-11-20 16:00:19.015711] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.868 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:20.868 "name": "raid_bdev1", 00:12:20.868 "uuid": "9515c950-90f0-4f20-82e5-8363dba52a0d", 00:12:20.868 "strip_size_kb": 0, 00:12:20.868 "state": "online", 00:12:20.868 "raid_level": "raid1", 00:12:20.868 "superblock": false, 00:12:20.868 "num_base_bdevs": 2, 00:12:20.868 "num_base_bdevs_discovered": 2, 00:12:20.868 "num_base_bdevs_operational": 2, 00:12:20.868 "base_bdevs_list": [ 00:12:20.868 { 00:12:20.868 "name": "BaseBdev1", 00:12:20.869 "uuid": "4a0c9364-b767-5513-865b-1c7175fdc326", 00:12:20.869 "is_configured": true, 00:12:20.869 "data_offset": 0, 00:12:20.869 "data_size": 65536 00:12:20.869 }, 00:12:20.869 { 00:12:20.869 "name": "BaseBdev2", 00:12:20.869 "uuid": "4de5b32c-7da9-5d11-8e00-4d7967a53e8c", 00:12:20.869 "is_configured": true, 00:12:20.869 "data_offset": 0, 00:12:20.869 "data_size": 65536 00:12:20.869 } 00:12:20.869 ] 00:12:20.869 }' 00:12:20.869 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:20.869 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:21.127 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:21.127 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:21.127 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:21.127 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:21.127 [2024-11-20 16:00:19.345454] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:21.127 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:21.127 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:12:21.127 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:21.127 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:21.127 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:21.127 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:12:21.385 [2024-11-20 16:00:19.589266] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:12:21.385 /dev/nbd0 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:21.385 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:21.646 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:12:21.646 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:21.646 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:21.646 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:21.646 1+0 records in 00:12:21.646 1+0 records out 00:12:21.646 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000414373 s, 9.9 MB/s 00:12:21.646 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:21.646 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:12:21.646 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:21.646 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:21.646 16:00:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:12:21.646 16:00:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:21.646 16:00:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:21.646 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:12:21.646 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:12:21.646 16:00:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:12:29.778 65536+0 records in 00:12:29.778 65536+0 records out 00:12:29.778 33554432 bytes (34 MB, 32 MiB) copied, 8.06328 s, 4.2 MB/s 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:29.778 [2024-11-20 16:00:27.993564] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:29.778 16:00:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:29.778 [2024-11-20 16:00:28.001665] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:29.778 16:00:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:29.778 16:00:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:29.778 16:00:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:29.778 16:00:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:29.778 16:00:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:29.778 16:00:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:29.778 16:00:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:29.778 16:00:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:29.778 16:00:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:29.778 16:00:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:29.778 16:00:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:29.778 16:00:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.778 16:00:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:29.778 16:00:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:29.778 16:00:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:30.037 16:00:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:30.037 16:00:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:30.037 "name": "raid_bdev1", 00:12:30.037 "uuid": "9515c950-90f0-4f20-82e5-8363dba52a0d", 00:12:30.037 "strip_size_kb": 0, 00:12:30.037 "state": "online", 00:12:30.037 "raid_level": "raid1", 00:12:30.037 "superblock": false, 00:12:30.037 "num_base_bdevs": 2, 00:12:30.037 "num_base_bdevs_discovered": 1, 00:12:30.037 "num_base_bdevs_operational": 1, 00:12:30.037 "base_bdevs_list": [ 00:12:30.037 { 00:12:30.037 "name": null, 00:12:30.037 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:30.037 "is_configured": false, 00:12:30.037 "data_offset": 0, 00:12:30.038 "data_size": 65536 00:12:30.038 }, 00:12:30.038 { 00:12:30.038 "name": "BaseBdev2", 00:12:30.038 "uuid": "4de5b32c-7da9-5d11-8e00-4d7967a53e8c", 00:12:30.038 "is_configured": true, 00:12:30.038 "data_offset": 0, 00:12:30.038 "data_size": 65536 00:12:30.038 } 00:12:30.038 ] 00:12:30.038 }' 00:12:30.038 16:00:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:30.038 16:00:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:30.296 16:00:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:30.296 16:00:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:30.296 16:00:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:30.296 [2024-11-20 16:00:28.325774] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:30.296 [2024-11-20 16:00:28.337469] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09bd0 00:12:30.296 16:00:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:30.296 16:00:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:30.296 [2024-11-20 16:00:28.339411] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:31.230 "name": "raid_bdev1", 00:12:31.230 "uuid": "9515c950-90f0-4f20-82e5-8363dba52a0d", 00:12:31.230 "strip_size_kb": 0, 00:12:31.230 "state": "online", 00:12:31.230 "raid_level": "raid1", 00:12:31.230 "superblock": false, 00:12:31.230 "num_base_bdevs": 2, 00:12:31.230 "num_base_bdevs_discovered": 2, 00:12:31.230 "num_base_bdevs_operational": 2, 00:12:31.230 "process": { 00:12:31.230 "type": "rebuild", 00:12:31.230 "target": "spare", 00:12:31.230 "progress": { 00:12:31.230 "blocks": 20480, 00:12:31.230 "percent": 31 00:12:31.230 } 00:12:31.230 }, 00:12:31.230 "base_bdevs_list": [ 00:12:31.230 { 00:12:31.230 "name": "spare", 00:12:31.230 "uuid": "259a7a9c-77e0-5633-b27f-cef15bb384bb", 00:12:31.230 "is_configured": true, 00:12:31.230 "data_offset": 0, 00:12:31.230 "data_size": 65536 00:12:31.230 }, 00:12:31.230 { 00:12:31.230 "name": "BaseBdev2", 00:12:31.230 "uuid": "4de5b32c-7da9-5d11-8e00-4d7967a53e8c", 00:12:31.230 "is_configured": true, 00:12:31.230 "data_offset": 0, 00:12:31.230 "data_size": 65536 00:12:31.230 } 00:12:31.230 ] 00:12:31.230 }' 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:31.230 [2024-11-20 16:00:29.429173] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:31.230 [2024-11-20 16:00:29.445176] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:31.230 [2024-11-20 16:00:29.445255] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:31.230 [2024-11-20 16:00:29.445270] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:31.230 [2024-11-20 16:00:29.445279] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:31.230 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:31.231 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:31.231 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:31.231 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:31.231 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:31.231 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:31.231 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:31.231 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:31.231 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.231 16:00:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.231 16:00:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:31.491 16:00:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.491 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:31.491 "name": "raid_bdev1", 00:12:31.491 "uuid": "9515c950-90f0-4f20-82e5-8363dba52a0d", 00:12:31.491 "strip_size_kb": 0, 00:12:31.491 "state": "online", 00:12:31.491 "raid_level": "raid1", 00:12:31.491 "superblock": false, 00:12:31.491 "num_base_bdevs": 2, 00:12:31.491 "num_base_bdevs_discovered": 1, 00:12:31.491 "num_base_bdevs_operational": 1, 00:12:31.491 "base_bdevs_list": [ 00:12:31.491 { 00:12:31.491 "name": null, 00:12:31.491 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.491 "is_configured": false, 00:12:31.491 "data_offset": 0, 00:12:31.491 "data_size": 65536 00:12:31.491 }, 00:12:31.491 { 00:12:31.491 "name": "BaseBdev2", 00:12:31.491 "uuid": "4de5b32c-7da9-5d11-8e00-4d7967a53e8c", 00:12:31.491 "is_configured": true, 00:12:31.491 "data_offset": 0, 00:12:31.491 "data_size": 65536 00:12:31.491 } 00:12:31.491 ] 00:12:31.491 }' 00:12:31.491 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:31.491 16:00:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:31.749 "name": "raid_bdev1", 00:12:31.749 "uuid": "9515c950-90f0-4f20-82e5-8363dba52a0d", 00:12:31.749 "strip_size_kb": 0, 00:12:31.749 "state": "online", 00:12:31.749 "raid_level": "raid1", 00:12:31.749 "superblock": false, 00:12:31.749 "num_base_bdevs": 2, 00:12:31.749 "num_base_bdevs_discovered": 1, 00:12:31.749 "num_base_bdevs_operational": 1, 00:12:31.749 "base_bdevs_list": [ 00:12:31.749 { 00:12:31.749 "name": null, 00:12:31.749 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.749 "is_configured": false, 00:12:31.749 "data_offset": 0, 00:12:31.749 "data_size": 65536 00:12:31.749 }, 00:12:31.749 { 00:12:31.749 "name": "BaseBdev2", 00:12:31.749 "uuid": "4de5b32c-7da9-5d11-8e00-4d7967a53e8c", 00:12:31.749 "is_configured": true, 00:12:31.749 "data_offset": 0, 00:12:31.749 "data_size": 65536 00:12:31.749 } 00:12:31.749 ] 00:12:31.749 }' 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:31.749 [2024-11-20 16:00:29.936769] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:31.749 [2024-11-20 16:00:29.948069] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09ca0 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.749 16:00:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:31.749 [2024-11-20 16:00:29.950080] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:33.126 16:00:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:33.126 16:00:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:33.126 16:00:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:33.126 16:00:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:33.126 16:00:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:33.126 16:00:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:33.126 16:00:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:33.126 16:00:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:33.127 16:00:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:33.127 16:00:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:33.127 16:00:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:33.127 "name": "raid_bdev1", 00:12:33.127 "uuid": "9515c950-90f0-4f20-82e5-8363dba52a0d", 00:12:33.127 "strip_size_kb": 0, 00:12:33.127 "state": "online", 00:12:33.127 "raid_level": "raid1", 00:12:33.127 "superblock": false, 00:12:33.127 "num_base_bdevs": 2, 00:12:33.127 "num_base_bdevs_discovered": 2, 00:12:33.127 "num_base_bdevs_operational": 2, 00:12:33.127 "process": { 00:12:33.127 "type": "rebuild", 00:12:33.127 "target": "spare", 00:12:33.127 "progress": { 00:12:33.127 "blocks": 20480, 00:12:33.127 "percent": 31 00:12:33.127 } 00:12:33.127 }, 00:12:33.127 "base_bdevs_list": [ 00:12:33.127 { 00:12:33.127 "name": "spare", 00:12:33.127 "uuid": "259a7a9c-77e0-5633-b27f-cef15bb384bb", 00:12:33.127 "is_configured": true, 00:12:33.127 "data_offset": 0, 00:12:33.127 "data_size": 65536 00:12:33.127 }, 00:12:33.127 { 00:12:33.127 "name": "BaseBdev2", 00:12:33.127 "uuid": "4de5b32c-7da9-5d11-8e00-4d7967a53e8c", 00:12:33.127 "is_configured": true, 00:12:33.127 "data_offset": 0, 00:12:33.127 "data_size": 65536 00:12:33.127 } 00:12:33.127 ] 00:12:33.127 }' 00:12:33.127 16:00:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=285 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:33.127 "name": "raid_bdev1", 00:12:33.127 "uuid": "9515c950-90f0-4f20-82e5-8363dba52a0d", 00:12:33.127 "strip_size_kb": 0, 00:12:33.127 "state": "online", 00:12:33.127 "raid_level": "raid1", 00:12:33.127 "superblock": false, 00:12:33.127 "num_base_bdevs": 2, 00:12:33.127 "num_base_bdevs_discovered": 2, 00:12:33.127 "num_base_bdevs_operational": 2, 00:12:33.127 "process": { 00:12:33.127 "type": "rebuild", 00:12:33.127 "target": "spare", 00:12:33.127 "progress": { 00:12:33.127 "blocks": 22528, 00:12:33.127 "percent": 34 00:12:33.127 } 00:12:33.127 }, 00:12:33.127 "base_bdevs_list": [ 00:12:33.127 { 00:12:33.127 "name": "spare", 00:12:33.127 "uuid": "259a7a9c-77e0-5633-b27f-cef15bb384bb", 00:12:33.127 "is_configured": true, 00:12:33.127 "data_offset": 0, 00:12:33.127 "data_size": 65536 00:12:33.127 }, 00:12:33.127 { 00:12:33.127 "name": "BaseBdev2", 00:12:33.127 "uuid": "4de5b32c-7da9-5d11-8e00-4d7967a53e8c", 00:12:33.127 "is_configured": true, 00:12:33.127 "data_offset": 0, 00:12:33.127 "data_size": 65536 00:12:33.127 } 00:12:33.127 ] 00:12:33.127 }' 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:33.127 16:00:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:34.068 16:00:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:34.068 16:00:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:34.068 16:00:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:34.068 16:00:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:34.068 16:00:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:34.068 16:00:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:34.068 16:00:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:34.068 16:00:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:34.068 16:00:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:34.068 16:00:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:34.068 16:00:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:34.068 16:00:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:34.068 "name": "raid_bdev1", 00:12:34.068 "uuid": "9515c950-90f0-4f20-82e5-8363dba52a0d", 00:12:34.068 "strip_size_kb": 0, 00:12:34.068 "state": "online", 00:12:34.068 "raid_level": "raid1", 00:12:34.068 "superblock": false, 00:12:34.068 "num_base_bdevs": 2, 00:12:34.068 "num_base_bdevs_discovered": 2, 00:12:34.068 "num_base_bdevs_operational": 2, 00:12:34.068 "process": { 00:12:34.068 "type": "rebuild", 00:12:34.068 "target": "spare", 00:12:34.068 "progress": { 00:12:34.068 "blocks": 45056, 00:12:34.068 "percent": 68 00:12:34.068 } 00:12:34.068 }, 00:12:34.068 "base_bdevs_list": [ 00:12:34.068 { 00:12:34.068 "name": "spare", 00:12:34.068 "uuid": "259a7a9c-77e0-5633-b27f-cef15bb384bb", 00:12:34.068 "is_configured": true, 00:12:34.068 "data_offset": 0, 00:12:34.068 "data_size": 65536 00:12:34.068 }, 00:12:34.068 { 00:12:34.068 "name": "BaseBdev2", 00:12:34.068 "uuid": "4de5b32c-7da9-5d11-8e00-4d7967a53e8c", 00:12:34.068 "is_configured": true, 00:12:34.068 "data_offset": 0, 00:12:34.068 "data_size": 65536 00:12:34.068 } 00:12:34.068 ] 00:12:34.068 }' 00:12:34.068 16:00:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:34.068 16:00:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:34.068 16:00:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:34.068 16:00:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:34.068 16:00:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:35.011 [2024-11-20 16:00:33.166759] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:35.011 [2024-11-20 16:00:33.166844] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:35.011 [2024-11-20 16:00:33.166891] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:35.271 "name": "raid_bdev1", 00:12:35.271 "uuid": "9515c950-90f0-4f20-82e5-8363dba52a0d", 00:12:35.271 "strip_size_kb": 0, 00:12:35.271 "state": "online", 00:12:35.271 "raid_level": "raid1", 00:12:35.271 "superblock": false, 00:12:35.271 "num_base_bdevs": 2, 00:12:35.271 "num_base_bdevs_discovered": 2, 00:12:35.271 "num_base_bdevs_operational": 2, 00:12:35.271 "base_bdevs_list": [ 00:12:35.271 { 00:12:35.271 "name": "spare", 00:12:35.271 "uuid": "259a7a9c-77e0-5633-b27f-cef15bb384bb", 00:12:35.271 "is_configured": true, 00:12:35.271 "data_offset": 0, 00:12:35.271 "data_size": 65536 00:12:35.271 }, 00:12:35.271 { 00:12:35.271 "name": "BaseBdev2", 00:12:35.271 "uuid": "4de5b32c-7da9-5d11-8e00-4d7967a53e8c", 00:12:35.271 "is_configured": true, 00:12:35.271 "data_offset": 0, 00:12:35.271 "data_size": 65536 00:12:35.271 } 00:12:35.271 ] 00:12:35.271 }' 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:35.271 "name": "raid_bdev1", 00:12:35.271 "uuid": "9515c950-90f0-4f20-82e5-8363dba52a0d", 00:12:35.271 "strip_size_kb": 0, 00:12:35.271 "state": "online", 00:12:35.271 "raid_level": "raid1", 00:12:35.271 "superblock": false, 00:12:35.271 "num_base_bdevs": 2, 00:12:35.271 "num_base_bdevs_discovered": 2, 00:12:35.271 "num_base_bdevs_operational": 2, 00:12:35.271 "base_bdevs_list": [ 00:12:35.271 { 00:12:35.271 "name": "spare", 00:12:35.271 "uuid": "259a7a9c-77e0-5633-b27f-cef15bb384bb", 00:12:35.271 "is_configured": true, 00:12:35.271 "data_offset": 0, 00:12:35.271 "data_size": 65536 00:12:35.271 }, 00:12:35.271 { 00:12:35.271 "name": "BaseBdev2", 00:12:35.271 "uuid": "4de5b32c-7da9-5d11-8e00-4d7967a53e8c", 00:12:35.271 "is_configured": true, 00:12:35.271 "data_offset": 0, 00:12:35.271 "data_size": 65536 00:12:35.271 } 00:12:35.271 ] 00:12:35.271 }' 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:35.271 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:35.272 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:35.272 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:35.272 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:35.272 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:35.272 16:00:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.272 16:00:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.272 16:00:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.272 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:35.272 "name": "raid_bdev1", 00:12:35.272 "uuid": "9515c950-90f0-4f20-82e5-8363dba52a0d", 00:12:35.272 "strip_size_kb": 0, 00:12:35.272 "state": "online", 00:12:35.272 "raid_level": "raid1", 00:12:35.272 "superblock": false, 00:12:35.272 "num_base_bdevs": 2, 00:12:35.272 "num_base_bdevs_discovered": 2, 00:12:35.272 "num_base_bdevs_operational": 2, 00:12:35.272 "base_bdevs_list": [ 00:12:35.272 { 00:12:35.272 "name": "spare", 00:12:35.272 "uuid": "259a7a9c-77e0-5633-b27f-cef15bb384bb", 00:12:35.272 "is_configured": true, 00:12:35.272 "data_offset": 0, 00:12:35.272 "data_size": 65536 00:12:35.272 }, 00:12:35.272 { 00:12:35.272 "name": "BaseBdev2", 00:12:35.272 "uuid": "4de5b32c-7da9-5d11-8e00-4d7967a53e8c", 00:12:35.272 "is_configured": true, 00:12:35.272 "data_offset": 0, 00:12:35.272 "data_size": 65536 00:12:35.272 } 00:12:35.272 ] 00:12:35.272 }' 00:12:35.272 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:35.272 16:00:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.615 [2024-11-20 16:00:33.794251] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:35.615 [2024-11-20 16:00:33.794466] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:35.615 [2024-11-20 16:00:33.794556] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:35.615 [2024-11-20 16:00:33.794629] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:35.615 [2024-11-20 16:00:33.794641] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:35.615 16:00:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:12:35.876 /dev/nbd0 00:12:35.876 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:35.876 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:35.876 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:35.876 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:12:35.876 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:35.876 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:35.876 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:35.876 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:12:35.876 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:35.876 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:35.877 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:35.877 1+0 records in 00:12:35.877 1+0 records out 00:12:35.877 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000278217 s, 14.7 MB/s 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:12:36.136 /dev/nbd1 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:36.136 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:36.395 1+0 records in 00:12:36.395 1+0 records out 00:12:36.395 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000324902 s, 12.6 MB/s 00:12:36.395 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:36.395 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:12:36.395 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:36.395 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:36.395 16:00:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:12:36.395 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:36.395 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:36.395 16:00:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:36.395 16:00:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:12:36.395 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:36.395 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:36.395 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:36.395 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:12:36.395 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:36.395 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:36.654 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:36.654 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:36.654 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:36.654 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:36.654 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:36.654 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:36.654 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:36.654 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:36.654 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:36.654 16:00:34 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:36.913 16:00:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:36.913 16:00:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:36.913 16:00:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:36.913 16:00:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:36.913 16:00:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:36.913 16:00:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:36.913 16:00:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:36.913 16:00:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:36.914 16:00:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:12:36.914 16:00:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 73179 00:12:36.914 16:00:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 73179 ']' 00:12:36.914 16:00:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 73179 00:12:36.914 16:00:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:12:36.914 16:00:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:36.914 16:00:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73179 00:12:36.914 killing process with pid 73179 00:12:36.914 16:00:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:36.914 16:00:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:36.914 16:00:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73179' 00:12:36.914 16:00:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@973 -- # kill 73179 00:12:36.914 Received shutdown signal, test time was about 60.000000 seconds 00:12:36.914 00:12:36.914 Latency(us) 00:12:36.914 [2024-11-20T16:00:35.165Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:36.914 [2024-11-20T16:00:35.165Z] =================================================================================================================== 00:12:36.914 [2024-11-20T16:00:35.165Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:36.914 16:00:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@978 -- # wait 73179 00:12:36.914 [2024-11-20 16:00:35.042975] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:37.174 [2024-11-20 16:00:35.230995] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:37.745 16:00:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:12:37.745 00:12:37.745 real 0m18.011s 00:12:37.745 user 0m18.794s 00:12:37.745 sys 0m3.831s 00:12:37.745 ************************************ 00:12:37.745 END TEST raid_rebuild_test 00:12:37.745 ************************************ 00:12:37.745 16:00:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:37.745 16:00:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.005 16:00:35 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 2 true false true 00:12:38.005 16:00:35 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:38.005 16:00:35 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:38.005 16:00:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:38.005 ************************************ 00:12:38.005 START TEST raid_rebuild_test_sb 00:12:38.005 ************************************ 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=73622 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 73622 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 73622 ']' 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:38.005 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:38.005 16:00:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:38.005 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:38.005 Zero copy mechanism will not be used. 00:12:38.005 [2024-11-20 16:00:36.084804] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:12:38.005 [2024-11-20 16:00:36.084927] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73622 ] 00:12:38.005 [2024-11-20 16:00:36.239944] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:38.265 [2024-11-20 16:00:36.348429] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:38.265 [2024-11-20 16:00:36.488452] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:38.265 [2024-11-20 16:00:36.488516] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:38.832 16:00:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:38.832 16:00:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:12:38.832 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:38.832 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:38.832 16:00:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:38.832 16:00:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:38.832 BaseBdev1_malloc 00:12:38.832 16:00:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:38.832 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:38.832 16:00:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:38.832 16:00:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:38.833 [2024-11-20 16:00:36.978292] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:38.833 [2024-11-20 16:00:36.978360] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:38.833 [2024-11-20 16:00:36.978383] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:38.833 [2024-11-20 16:00:36.978394] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:38.833 [2024-11-20 16:00:36.980651] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:38.833 [2024-11-20 16:00:36.980699] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:38.833 BaseBdev1 00:12:38.833 16:00:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:38.833 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:38.833 16:00:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:38.833 16:00:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:38.833 16:00:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:38.833 BaseBdev2_malloc 00:12:38.833 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:38.833 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:38.833 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:38.833 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:38.833 [2024-11-20 16:00:37.019013] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:38.833 [2024-11-20 16:00:37.019098] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:38.833 [2024-11-20 16:00:37.019121] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:38.833 [2024-11-20 16:00:37.019132] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:38.833 [2024-11-20 16:00:37.021311] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:38.833 [2024-11-20 16:00:37.021346] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:38.833 BaseBdev2 00:12:38.833 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:38.833 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:38.833 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:38.833 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:38.833 spare_malloc 00:12:38.833 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:38.833 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:38.833 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:38.833 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:38.833 spare_delay 00:12:38.833 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:38.833 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:38.833 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:38.833 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:38.833 [2024-11-20 16:00:37.079146] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:38.833 [2024-11-20 16:00:37.079202] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:38.833 [2024-11-20 16:00:37.079224] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:12:38.833 [2024-11-20 16:00:37.079234] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:39.091 [2024-11-20 16:00:37.081420] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:39.091 [2024-11-20 16:00:37.081454] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:39.091 spare 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:39.091 [2024-11-20 16:00:37.087197] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:39.091 [2024-11-20 16:00:37.089062] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:39.091 [2024-11-20 16:00:37.089238] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:12:39.091 [2024-11-20 16:00:37.089251] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:39.091 [2024-11-20 16:00:37.089515] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:12:39.091 [2024-11-20 16:00:37.089687] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:12:39.091 [2024-11-20 16:00:37.089697] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:12:39.091 [2024-11-20 16:00:37.089854] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:39.091 "name": "raid_bdev1", 00:12:39.091 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:39.091 "strip_size_kb": 0, 00:12:39.091 "state": "online", 00:12:39.091 "raid_level": "raid1", 00:12:39.091 "superblock": true, 00:12:39.091 "num_base_bdevs": 2, 00:12:39.091 "num_base_bdevs_discovered": 2, 00:12:39.091 "num_base_bdevs_operational": 2, 00:12:39.091 "base_bdevs_list": [ 00:12:39.091 { 00:12:39.091 "name": "BaseBdev1", 00:12:39.091 "uuid": "6fcdba71-f992-5ef2-9dd7-882275838a2f", 00:12:39.091 "is_configured": true, 00:12:39.091 "data_offset": 2048, 00:12:39.091 "data_size": 63488 00:12:39.091 }, 00:12:39.091 { 00:12:39.091 "name": "BaseBdev2", 00:12:39.091 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:39.091 "is_configured": true, 00:12:39.091 "data_offset": 2048, 00:12:39.091 "data_size": 63488 00:12:39.091 } 00:12:39.091 ] 00:12:39.091 }' 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:39.091 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:39.351 [2024-11-20 16:00:37.455559] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:39.351 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:12:39.610 [2024-11-20 16:00:37.699367] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:12:39.610 /dev/nbd0 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:39.610 1+0 records in 00:12:39.610 1+0 records out 00:12:39.610 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000492184 s, 8.3 MB/s 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:12:39.610 16:00:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:12:46.281 63488+0 records in 00:12:46.281 63488+0 records out 00:12:46.281 32505856 bytes (33 MB, 31 MiB) copied, 6.58905 s, 4.9 MB/s 00:12:46.281 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:46.281 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:46.281 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:46.281 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:46.281 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:12:46.281 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:46.281 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:46.563 [2024-11-20 16:00:44.543688] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.563 [2024-11-20 16:00:44.579771] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.563 16:00:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.564 16:00:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.564 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:46.564 "name": "raid_bdev1", 00:12:46.564 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:46.564 "strip_size_kb": 0, 00:12:46.564 "state": "online", 00:12:46.564 "raid_level": "raid1", 00:12:46.564 "superblock": true, 00:12:46.564 "num_base_bdevs": 2, 00:12:46.564 "num_base_bdevs_discovered": 1, 00:12:46.564 "num_base_bdevs_operational": 1, 00:12:46.564 "base_bdevs_list": [ 00:12:46.564 { 00:12:46.564 "name": null, 00:12:46.564 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:46.564 "is_configured": false, 00:12:46.564 "data_offset": 0, 00:12:46.564 "data_size": 63488 00:12:46.564 }, 00:12:46.564 { 00:12:46.564 "name": "BaseBdev2", 00:12:46.564 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:46.564 "is_configured": true, 00:12:46.564 "data_offset": 2048, 00:12:46.564 "data_size": 63488 00:12:46.564 } 00:12:46.564 ] 00:12:46.564 }' 00:12:46.564 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:46.564 16:00:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.823 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:46.823 16:00:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.823 16:00:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.823 [2024-11-20 16:00:44.883842] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:46.823 [2024-11-20 16:00:44.895475] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca3360 00:12:46.823 16:00:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.823 16:00:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:46.823 [2024-11-20 16:00:44.897452] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:47.796 16:00:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:47.796 16:00:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:47.796 16:00:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:47.796 16:00:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:47.796 16:00:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:47.796 16:00:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:47.796 16:00:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.796 16:00:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.796 16:00:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.796 16:00:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.796 16:00:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:47.796 "name": "raid_bdev1", 00:12:47.796 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:47.796 "strip_size_kb": 0, 00:12:47.796 "state": "online", 00:12:47.796 "raid_level": "raid1", 00:12:47.796 "superblock": true, 00:12:47.796 "num_base_bdevs": 2, 00:12:47.796 "num_base_bdevs_discovered": 2, 00:12:47.796 "num_base_bdevs_operational": 2, 00:12:47.796 "process": { 00:12:47.796 "type": "rebuild", 00:12:47.796 "target": "spare", 00:12:47.796 "progress": { 00:12:47.796 "blocks": 20480, 00:12:47.796 "percent": 32 00:12:47.796 } 00:12:47.796 }, 00:12:47.796 "base_bdevs_list": [ 00:12:47.796 { 00:12:47.796 "name": "spare", 00:12:47.796 "uuid": "51bc3159-85ff-5813-96d7-64d4f2273c59", 00:12:47.796 "is_configured": true, 00:12:47.796 "data_offset": 2048, 00:12:47.796 "data_size": 63488 00:12:47.796 }, 00:12:47.796 { 00:12:47.796 "name": "BaseBdev2", 00:12:47.796 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:47.796 "is_configured": true, 00:12:47.796 "data_offset": 2048, 00:12:47.796 "data_size": 63488 00:12:47.796 } 00:12:47.796 ] 00:12:47.796 }' 00:12:47.796 16:00:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:47.796 16:00:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:47.796 16:00:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:47.796 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:47.796 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:47.796 16:00:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.796 16:00:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.796 [2024-11-20 16:00:46.007571] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:48.054 [2024-11-20 16:00:46.103628] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:48.054 [2024-11-20 16:00:46.103722] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:48.054 [2024-11-20 16:00:46.103748] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:48.054 [2024-11-20 16:00:46.103761] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:48.054 16:00:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.054 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:48.054 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:48.054 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:48.054 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:48.054 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:48.054 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:48.054 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:48.054 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:48.054 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:48.054 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:48.054 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.054 16:00:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.054 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:48.054 16:00:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.054 16:00:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.054 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:48.054 "name": "raid_bdev1", 00:12:48.054 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:48.054 "strip_size_kb": 0, 00:12:48.054 "state": "online", 00:12:48.054 "raid_level": "raid1", 00:12:48.054 "superblock": true, 00:12:48.054 "num_base_bdevs": 2, 00:12:48.054 "num_base_bdevs_discovered": 1, 00:12:48.054 "num_base_bdevs_operational": 1, 00:12:48.054 "base_bdevs_list": [ 00:12:48.054 { 00:12:48.054 "name": null, 00:12:48.054 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:48.054 "is_configured": false, 00:12:48.054 "data_offset": 0, 00:12:48.054 "data_size": 63488 00:12:48.054 }, 00:12:48.054 { 00:12:48.054 "name": "BaseBdev2", 00:12:48.054 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:48.054 "is_configured": true, 00:12:48.054 "data_offset": 2048, 00:12:48.055 "data_size": 63488 00:12:48.055 } 00:12:48.055 ] 00:12:48.055 }' 00:12:48.055 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:48.055 16:00:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.313 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:48.313 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:48.313 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:48.313 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:48.313 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:48.314 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.314 16:00:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.314 16:00:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.314 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:48.314 16:00:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.314 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:48.314 "name": "raid_bdev1", 00:12:48.314 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:48.314 "strip_size_kb": 0, 00:12:48.314 "state": "online", 00:12:48.314 "raid_level": "raid1", 00:12:48.314 "superblock": true, 00:12:48.314 "num_base_bdevs": 2, 00:12:48.314 "num_base_bdevs_discovered": 1, 00:12:48.314 "num_base_bdevs_operational": 1, 00:12:48.314 "base_bdevs_list": [ 00:12:48.314 { 00:12:48.314 "name": null, 00:12:48.314 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:48.314 "is_configured": false, 00:12:48.314 "data_offset": 0, 00:12:48.314 "data_size": 63488 00:12:48.314 }, 00:12:48.314 { 00:12:48.314 "name": "BaseBdev2", 00:12:48.314 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:48.314 "is_configured": true, 00:12:48.314 "data_offset": 2048, 00:12:48.314 "data_size": 63488 00:12:48.314 } 00:12:48.314 ] 00:12:48.314 }' 00:12:48.314 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:48.314 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:48.314 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:48.314 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:48.314 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:48.314 16:00:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.314 16:00:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.571 [2024-11-20 16:00:46.563091] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:48.571 [2024-11-20 16:00:46.573874] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca3430 00:12:48.571 16:00:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.571 16:00:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:48.571 [2024-11-20 16:00:46.575914] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:49.504 "name": "raid_bdev1", 00:12:49.504 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:49.504 "strip_size_kb": 0, 00:12:49.504 "state": "online", 00:12:49.504 "raid_level": "raid1", 00:12:49.504 "superblock": true, 00:12:49.504 "num_base_bdevs": 2, 00:12:49.504 "num_base_bdevs_discovered": 2, 00:12:49.504 "num_base_bdevs_operational": 2, 00:12:49.504 "process": { 00:12:49.504 "type": "rebuild", 00:12:49.504 "target": "spare", 00:12:49.504 "progress": { 00:12:49.504 "blocks": 20480, 00:12:49.504 "percent": 32 00:12:49.504 } 00:12:49.504 }, 00:12:49.504 "base_bdevs_list": [ 00:12:49.504 { 00:12:49.504 "name": "spare", 00:12:49.504 "uuid": "51bc3159-85ff-5813-96d7-64d4f2273c59", 00:12:49.504 "is_configured": true, 00:12:49.504 "data_offset": 2048, 00:12:49.504 "data_size": 63488 00:12:49.504 }, 00:12:49.504 { 00:12:49.504 "name": "BaseBdev2", 00:12:49.504 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:49.504 "is_configured": true, 00:12:49.504 "data_offset": 2048, 00:12:49.504 "data_size": 63488 00:12:49.504 } 00:12:49.504 ] 00:12:49.504 }' 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:12:49.504 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=301 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.504 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:49.504 "name": "raid_bdev1", 00:12:49.505 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:49.505 "strip_size_kb": 0, 00:12:49.505 "state": "online", 00:12:49.505 "raid_level": "raid1", 00:12:49.505 "superblock": true, 00:12:49.505 "num_base_bdevs": 2, 00:12:49.505 "num_base_bdevs_discovered": 2, 00:12:49.505 "num_base_bdevs_operational": 2, 00:12:49.505 "process": { 00:12:49.505 "type": "rebuild", 00:12:49.505 "target": "spare", 00:12:49.505 "progress": { 00:12:49.505 "blocks": 22528, 00:12:49.505 "percent": 35 00:12:49.505 } 00:12:49.505 }, 00:12:49.505 "base_bdevs_list": [ 00:12:49.505 { 00:12:49.505 "name": "spare", 00:12:49.505 "uuid": "51bc3159-85ff-5813-96d7-64d4f2273c59", 00:12:49.505 "is_configured": true, 00:12:49.505 "data_offset": 2048, 00:12:49.505 "data_size": 63488 00:12:49.505 }, 00:12:49.505 { 00:12:49.505 "name": "BaseBdev2", 00:12:49.505 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:49.505 "is_configured": true, 00:12:49.505 "data_offset": 2048, 00:12:49.505 "data_size": 63488 00:12:49.505 } 00:12:49.505 ] 00:12:49.505 }' 00:12:49.505 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:49.505 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:49.505 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:49.762 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:49.762 16:00:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:50.694 16:00:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:50.694 16:00:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:50.694 16:00:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:50.694 16:00:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:50.694 16:00:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:50.694 16:00:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:50.694 16:00:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:50.694 16:00:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:50.694 16:00:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:50.694 16:00:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.694 16:00:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:50.694 16:00:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:50.694 "name": "raid_bdev1", 00:12:50.694 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:50.694 "strip_size_kb": 0, 00:12:50.694 "state": "online", 00:12:50.694 "raid_level": "raid1", 00:12:50.694 "superblock": true, 00:12:50.694 "num_base_bdevs": 2, 00:12:50.694 "num_base_bdevs_discovered": 2, 00:12:50.694 "num_base_bdevs_operational": 2, 00:12:50.694 "process": { 00:12:50.694 "type": "rebuild", 00:12:50.694 "target": "spare", 00:12:50.694 "progress": { 00:12:50.694 "blocks": 45056, 00:12:50.694 "percent": 70 00:12:50.694 } 00:12:50.694 }, 00:12:50.694 "base_bdevs_list": [ 00:12:50.694 { 00:12:50.694 "name": "spare", 00:12:50.694 "uuid": "51bc3159-85ff-5813-96d7-64d4f2273c59", 00:12:50.694 "is_configured": true, 00:12:50.694 "data_offset": 2048, 00:12:50.694 "data_size": 63488 00:12:50.694 }, 00:12:50.694 { 00:12:50.694 "name": "BaseBdev2", 00:12:50.694 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:50.694 "is_configured": true, 00:12:50.694 "data_offset": 2048, 00:12:50.694 "data_size": 63488 00:12:50.694 } 00:12:50.694 ] 00:12:50.694 }' 00:12:50.694 16:00:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:50.694 16:00:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:50.694 16:00:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:50.694 16:00:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:50.694 16:00:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:51.628 [2024-11-20 16:00:49.690526] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:51.628 [2024-11-20 16:00:49.690593] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:51.628 [2024-11-20 16:00:49.690707] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:51.886 16:00:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:51.886 16:00:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:51.886 16:00:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:51.886 16:00:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:51.886 16:00:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:51.886 16:00:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:51.886 16:00:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:51.886 16:00:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:51.886 16:00:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.886 16:00:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.886 16:00:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.886 16:00:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:51.886 "name": "raid_bdev1", 00:12:51.886 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:51.886 "strip_size_kb": 0, 00:12:51.886 "state": "online", 00:12:51.886 "raid_level": "raid1", 00:12:51.886 "superblock": true, 00:12:51.886 "num_base_bdevs": 2, 00:12:51.886 "num_base_bdevs_discovered": 2, 00:12:51.886 "num_base_bdevs_operational": 2, 00:12:51.886 "base_bdevs_list": [ 00:12:51.886 { 00:12:51.886 "name": "spare", 00:12:51.886 "uuid": "51bc3159-85ff-5813-96d7-64d4f2273c59", 00:12:51.886 "is_configured": true, 00:12:51.886 "data_offset": 2048, 00:12:51.886 "data_size": 63488 00:12:51.886 }, 00:12:51.886 { 00:12:51.886 "name": "BaseBdev2", 00:12:51.886 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:51.886 "is_configured": true, 00:12:51.886 "data_offset": 2048, 00:12:51.886 "data_size": 63488 00:12:51.886 } 00:12:51.886 ] 00:12:51.886 }' 00:12:51.886 16:00:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:51.886 16:00:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:51.886 16:00:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:51.886 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:51.886 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:12:51.886 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:51.886 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:51.886 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:51.886 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:51.886 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:51.886 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:51.886 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.886 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.886 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:51.886 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.886 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:51.886 "name": "raid_bdev1", 00:12:51.886 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:51.886 "strip_size_kb": 0, 00:12:51.886 "state": "online", 00:12:51.886 "raid_level": "raid1", 00:12:51.886 "superblock": true, 00:12:51.886 "num_base_bdevs": 2, 00:12:51.886 "num_base_bdevs_discovered": 2, 00:12:51.886 "num_base_bdevs_operational": 2, 00:12:51.886 "base_bdevs_list": [ 00:12:51.886 { 00:12:51.886 "name": "spare", 00:12:51.886 "uuid": "51bc3159-85ff-5813-96d7-64d4f2273c59", 00:12:51.886 "is_configured": true, 00:12:51.887 "data_offset": 2048, 00:12:51.887 "data_size": 63488 00:12:51.887 }, 00:12:51.887 { 00:12:51.887 "name": "BaseBdev2", 00:12:51.887 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:51.887 "is_configured": true, 00:12:51.887 "data_offset": 2048, 00:12:51.887 "data_size": 63488 00:12:51.887 } 00:12:51.887 ] 00:12:51.887 }' 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:51.887 "name": "raid_bdev1", 00:12:51.887 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:51.887 "strip_size_kb": 0, 00:12:51.887 "state": "online", 00:12:51.887 "raid_level": "raid1", 00:12:51.887 "superblock": true, 00:12:51.887 "num_base_bdevs": 2, 00:12:51.887 "num_base_bdevs_discovered": 2, 00:12:51.887 "num_base_bdevs_operational": 2, 00:12:51.887 "base_bdevs_list": [ 00:12:51.887 { 00:12:51.887 "name": "spare", 00:12:51.887 "uuid": "51bc3159-85ff-5813-96d7-64d4f2273c59", 00:12:51.887 "is_configured": true, 00:12:51.887 "data_offset": 2048, 00:12:51.887 "data_size": 63488 00:12:51.887 }, 00:12:51.887 { 00:12:51.887 "name": "BaseBdev2", 00:12:51.887 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:51.887 "is_configured": true, 00:12:51.887 "data_offset": 2048, 00:12:51.887 "data_size": 63488 00:12:51.887 } 00:12:51.887 ] 00:12:51.887 }' 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:51.887 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:52.454 [2024-11-20 16:00:50.424976] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:52.454 [2024-11-20 16:00:50.425004] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:52.454 [2024-11-20 16:00:50.425063] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:52.454 [2024-11-20 16:00:50.425116] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:52.454 [2024-11-20 16:00:50.425125] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:12:52.454 /dev/nbd0 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:52.454 1+0 records in 00:12:52.454 1+0 records out 00:12:52.454 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000306261 s, 13.4 MB/s 00:12:52.454 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:12:52.711 /dev/nbd1 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:52.711 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:52.711 1+0 records in 00:12:52.712 1+0 records out 00:12:52.712 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000294107 s, 13.9 MB/s 00:12:52.712 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:52.712 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:12:52.712 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:52.712 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:52.712 16:00:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:12:52.712 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:52.712 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:52.712 16:00:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:52.970 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:12:52.970 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:52.970 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:52.970 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:52.970 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:12:52.970 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:52.970 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:53.228 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:53.228 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:53.228 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:53.228 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:53.228 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:53.228 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:53.228 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:53.228 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:53.228 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:53.228 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:53.228 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.486 [2024-11-20 16:00:51.497022] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:53.486 [2024-11-20 16:00:51.497077] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:53.486 [2024-11-20 16:00:51.497097] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:12:53.486 [2024-11-20 16:00:51.497105] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:53.486 [2024-11-20 16:00:51.499019] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:53.486 [2024-11-20 16:00:51.499050] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:53.486 [2024-11-20 16:00:51.499127] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:53.486 [2024-11-20 16:00:51.499164] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:53.486 [2024-11-20 16:00:51.499272] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:53.486 spare 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.486 [2024-11-20 16:00:51.599356] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:12:53.486 [2024-11-20 16:00:51.599404] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:53.486 [2024-11-20 16:00:51.599687] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1ae0 00:12:53.486 [2024-11-20 16:00:51.599846] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:12:53.486 [2024-11-20 16:00:51.599861] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:12:53.486 [2024-11-20 16:00:51.600002] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.486 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:53.486 "name": "raid_bdev1", 00:12:53.486 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:53.486 "strip_size_kb": 0, 00:12:53.486 "state": "online", 00:12:53.486 "raid_level": "raid1", 00:12:53.486 "superblock": true, 00:12:53.486 "num_base_bdevs": 2, 00:12:53.486 "num_base_bdevs_discovered": 2, 00:12:53.487 "num_base_bdevs_operational": 2, 00:12:53.487 "base_bdevs_list": [ 00:12:53.487 { 00:12:53.487 "name": "spare", 00:12:53.487 "uuid": "51bc3159-85ff-5813-96d7-64d4f2273c59", 00:12:53.487 "is_configured": true, 00:12:53.487 "data_offset": 2048, 00:12:53.487 "data_size": 63488 00:12:53.487 }, 00:12:53.487 { 00:12:53.487 "name": "BaseBdev2", 00:12:53.487 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:53.487 "is_configured": true, 00:12:53.487 "data_offset": 2048, 00:12:53.487 "data_size": 63488 00:12:53.487 } 00:12:53.487 ] 00:12:53.487 }' 00:12:53.487 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:53.487 16:00:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.745 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:53.745 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:53.745 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:53.745 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:53.745 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:53.745 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.745 16:00:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.745 16:00:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.745 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:53.745 16:00:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.745 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:53.745 "name": "raid_bdev1", 00:12:53.745 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:53.745 "strip_size_kb": 0, 00:12:53.745 "state": "online", 00:12:53.745 "raid_level": "raid1", 00:12:53.745 "superblock": true, 00:12:53.745 "num_base_bdevs": 2, 00:12:53.745 "num_base_bdevs_discovered": 2, 00:12:53.745 "num_base_bdevs_operational": 2, 00:12:53.745 "base_bdevs_list": [ 00:12:53.745 { 00:12:53.745 "name": "spare", 00:12:53.745 "uuid": "51bc3159-85ff-5813-96d7-64d4f2273c59", 00:12:53.745 "is_configured": true, 00:12:53.745 "data_offset": 2048, 00:12:53.745 "data_size": 63488 00:12:53.745 }, 00:12:53.745 { 00:12:53.745 "name": "BaseBdev2", 00:12:53.745 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:53.745 "is_configured": true, 00:12:53.745 "data_offset": 2048, 00:12:53.745 "data_size": 63488 00:12:53.745 } 00:12:53.745 ] 00:12:53.745 }' 00:12:53.745 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:53.745 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:53.745 16:00:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:54.003 [2024-11-20 16:00:52.061175] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:54.003 "name": "raid_bdev1", 00:12:54.003 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:54.003 "strip_size_kb": 0, 00:12:54.003 "state": "online", 00:12:54.003 "raid_level": "raid1", 00:12:54.003 "superblock": true, 00:12:54.003 "num_base_bdevs": 2, 00:12:54.003 "num_base_bdevs_discovered": 1, 00:12:54.003 "num_base_bdevs_operational": 1, 00:12:54.003 "base_bdevs_list": [ 00:12:54.003 { 00:12:54.003 "name": null, 00:12:54.003 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:54.003 "is_configured": false, 00:12:54.003 "data_offset": 0, 00:12:54.003 "data_size": 63488 00:12:54.003 }, 00:12:54.003 { 00:12:54.003 "name": "BaseBdev2", 00:12:54.003 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:54.003 "is_configured": true, 00:12:54.003 "data_offset": 2048, 00:12:54.003 "data_size": 63488 00:12:54.003 } 00:12:54.003 ] 00:12:54.003 }' 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:54.003 16:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:54.261 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:54.261 16:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.261 16:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:54.261 [2024-11-20 16:00:52.397259] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:54.261 [2024-11-20 16:00:52.397417] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:12:54.261 [2024-11-20 16:00:52.397431] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:54.261 [2024-11-20 16:00:52.397464] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:54.261 [2024-11-20 16:00:52.406186] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1bb0 00:12:54.261 16:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.261 16:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:12:54.261 [2024-11-20 16:00:52.407802] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:55.194 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:55.194 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:55.194 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:55.194 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:55.194 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:55.194 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.194 16:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.194 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:55.194 16:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.194 16:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:55.452 "name": "raid_bdev1", 00:12:55.452 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:55.452 "strip_size_kb": 0, 00:12:55.452 "state": "online", 00:12:55.452 "raid_level": "raid1", 00:12:55.452 "superblock": true, 00:12:55.452 "num_base_bdevs": 2, 00:12:55.452 "num_base_bdevs_discovered": 2, 00:12:55.452 "num_base_bdevs_operational": 2, 00:12:55.452 "process": { 00:12:55.452 "type": "rebuild", 00:12:55.452 "target": "spare", 00:12:55.452 "progress": { 00:12:55.452 "blocks": 20480, 00:12:55.452 "percent": 32 00:12:55.452 } 00:12:55.452 }, 00:12:55.452 "base_bdevs_list": [ 00:12:55.452 { 00:12:55.452 "name": "spare", 00:12:55.452 "uuid": "51bc3159-85ff-5813-96d7-64d4f2273c59", 00:12:55.452 "is_configured": true, 00:12:55.452 "data_offset": 2048, 00:12:55.452 "data_size": 63488 00:12:55.452 }, 00:12:55.452 { 00:12:55.452 "name": "BaseBdev2", 00:12:55.452 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:55.452 "is_configured": true, 00:12:55.452 "data_offset": 2048, 00:12:55.452 "data_size": 63488 00:12:55.452 } 00:12:55.452 ] 00:12:55.452 }' 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.452 [2024-11-20 16:00:53.518121] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:55.452 [2024-11-20 16:00:53.613414] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:55.452 [2024-11-20 16:00:53.613481] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:55.452 [2024-11-20 16:00:53.613494] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:55.452 [2024-11-20 16:00:53.613502] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.452 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:55.452 "name": "raid_bdev1", 00:12:55.452 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:55.452 "strip_size_kb": 0, 00:12:55.452 "state": "online", 00:12:55.452 "raid_level": "raid1", 00:12:55.452 "superblock": true, 00:12:55.452 "num_base_bdevs": 2, 00:12:55.452 "num_base_bdevs_discovered": 1, 00:12:55.452 "num_base_bdevs_operational": 1, 00:12:55.453 "base_bdevs_list": [ 00:12:55.453 { 00:12:55.453 "name": null, 00:12:55.453 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:55.453 "is_configured": false, 00:12:55.453 "data_offset": 0, 00:12:55.453 "data_size": 63488 00:12:55.453 }, 00:12:55.453 { 00:12:55.453 "name": "BaseBdev2", 00:12:55.453 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:55.453 "is_configured": true, 00:12:55.453 "data_offset": 2048, 00:12:55.453 "data_size": 63488 00:12:55.453 } 00:12:55.453 ] 00:12:55.453 }' 00:12:55.453 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:55.453 16:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.710 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:55.710 16:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.710 16:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.710 [2024-11-20 16:00:53.952198] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:55.710 [2024-11-20 16:00:53.952257] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:55.710 [2024-11-20 16:00:53.952273] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:12:55.710 [2024-11-20 16:00:53.952282] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:55.710 [2024-11-20 16:00:53.952660] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:55.710 [2024-11-20 16:00:53.952694] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:55.710 [2024-11-20 16:00:53.952771] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:55.710 [2024-11-20 16:00:53.952784] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:12:55.710 [2024-11-20 16:00:53.952792] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:55.710 [2024-11-20 16:00:53.952813] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:55.967 [2024-11-20 16:00:53.961843] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1c80 00:12:55.967 spare 00:12:55.967 16:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.967 16:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:12:55.967 [2024-11-20 16:00:53.963456] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:56.900 16:00:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:56.900 16:00:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:56.900 16:00:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:56.900 16:00:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:56.900 16:00:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:56.900 16:00:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:56.900 16:00:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:56.900 16:00:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:56.900 16:00:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:56.900 16:00:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:56.900 16:00:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:56.900 "name": "raid_bdev1", 00:12:56.900 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:56.900 "strip_size_kb": 0, 00:12:56.900 "state": "online", 00:12:56.900 "raid_level": "raid1", 00:12:56.900 "superblock": true, 00:12:56.900 "num_base_bdevs": 2, 00:12:56.900 "num_base_bdevs_discovered": 2, 00:12:56.900 "num_base_bdevs_operational": 2, 00:12:56.900 "process": { 00:12:56.900 "type": "rebuild", 00:12:56.900 "target": "spare", 00:12:56.900 "progress": { 00:12:56.900 "blocks": 20480, 00:12:56.900 "percent": 32 00:12:56.900 } 00:12:56.900 }, 00:12:56.900 "base_bdevs_list": [ 00:12:56.900 { 00:12:56.900 "name": "spare", 00:12:56.900 "uuid": "51bc3159-85ff-5813-96d7-64d4f2273c59", 00:12:56.900 "is_configured": true, 00:12:56.900 "data_offset": 2048, 00:12:56.900 "data_size": 63488 00:12:56.900 }, 00:12:56.900 { 00:12:56.900 "name": "BaseBdev2", 00:12:56.900 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:56.900 "is_configured": true, 00:12:56.900 "data_offset": 2048, 00:12:56.900 "data_size": 63488 00:12:56.900 } 00:12:56.900 ] 00:12:56.900 }' 00:12:56.900 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:56.900 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:56.900 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:56.900 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:56.900 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:12:56.900 16:00:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:56.900 16:00:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:56.900 [2024-11-20 16:00:55.077473] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:57.158 [2024-11-20 16:00:55.169173] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:57.158 [2024-11-20 16:00:55.169245] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:57.158 [2024-11-20 16:00:55.169260] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:57.158 [2024-11-20 16:00:55.169266] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:57.158 "name": "raid_bdev1", 00:12:57.158 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:57.158 "strip_size_kb": 0, 00:12:57.158 "state": "online", 00:12:57.158 "raid_level": "raid1", 00:12:57.158 "superblock": true, 00:12:57.158 "num_base_bdevs": 2, 00:12:57.158 "num_base_bdevs_discovered": 1, 00:12:57.158 "num_base_bdevs_operational": 1, 00:12:57.158 "base_bdevs_list": [ 00:12:57.158 { 00:12:57.158 "name": null, 00:12:57.158 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:57.158 "is_configured": false, 00:12:57.158 "data_offset": 0, 00:12:57.158 "data_size": 63488 00:12:57.158 }, 00:12:57.158 { 00:12:57.158 "name": "BaseBdev2", 00:12:57.158 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:57.158 "is_configured": true, 00:12:57.158 "data_offset": 2048, 00:12:57.158 "data_size": 63488 00:12:57.158 } 00:12:57.158 ] 00:12:57.158 }' 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:57.158 16:00:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:57.416 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:57.416 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:57.416 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:57.416 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:57.416 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:57.416 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:57.416 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:57.416 16:00:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.417 16:00:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:57.417 16:00:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.417 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:57.417 "name": "raid_bdev1", 00:12:57.417 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:57.417 "strip_size_kb": 0, 00:12:57.417 "state": "online", 00:12:57.417 "raid_level": "raid1", 00:12:57.417 "superblock": true, 00:12:57.417 "num_base_bdevs": 2, 00:12:57.417 "num_base_bdevs_discovered": 1, 00:12:57.417 "num_base_bdevs_operational": 1, 00:12:57.417 "base_bdevs_list": [ 00:12:57.417 { 00:12:57.417 "name": null, 00:12:57.417 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:57.417 "is_configured": false, 00:12:57.417 "data_offset": 0, 00:12:57.417 "data_size": 63488 00:12:57.417 }, 00:12:57.417 { 00:12:57.417 "name": "BaseBdev2", 00:12:57.417 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:57.417 "is_configured": true, 00:12:57.417 "data_offset": 2048, 00:12:57.417 "data_size": 63488 00:12:57.417 } 00:12:57.417 ] 00:12:57.417 }' 00:12:57.417 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:57.417 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:57.417 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:57.417 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:57.417 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:12:57.417 16:00:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.417 16:00:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:57.417 16:00:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.417 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:57.417 16:00:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.417 16:00:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:57.417 [2024-11-20 16:00:55.620213] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:57.417 [2024-11-20 16:00:55.620267] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:57.417 [2024-11-20 16:00:55.620291] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:12:57.417 [2024-11-20 16:00:55.620304] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:57.417 [2024-11-20 16:00:55.620694] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:57.417 [2024-11-20 16:00:55.620713] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:57.417 [2024-11-20 16:00:55.620780] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:12:57.417 [2024-11-20 16:00:55.620791] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:12:57.417 [2024-11-20 16:00:55.620799] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:57.417 [2024-11-20 16:00:55.620807] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:12:57.417 BaseBdev1 00:12:57.417 16:00:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.417 16:00:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:12:58.419 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:58.419 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:58.419 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:58.419 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:58.419 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:58.419 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:58.419 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:58.419 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:58.419 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:58.419 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:58.419 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:58.419 16:00:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.419 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:58.419 16:00:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:58.419 16:00:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.677 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:58.677 "name": "raid_bdev1", 00:12:58.677 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:58.677 "strip_size_kb": 0, 00:12:58.677 "state": "online", 00:12:58.677 "raid_level": "raid1", 00:12:58.677 "superblock": true, 00:12:58.677 "num_base_bdevs": 2, 00:12:58.677 "num_base_bdevs_discovered": 1, 00:12:58.677 "num_base_bdevs_operational": 1, 00:12:58.677 "base_bdevs_list": [ 00:12:58.677 { 00:12:58.677 "name": null, 00:12:58.677 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:58.677 "is_configured": false, 00:12:58.677 "data_offset": 0, 00:12:58.677 "data_size": 63488 00:12:58.677 }, 00:12:58.677 { 00:12:58.677 "name": "BaseBdev2", 00:12:58.677 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:58.677 "is_configured": true, 00:12:58.677 "data_offset": 2048, 00:12:58.677 "data_size": 63488 00:12:58.677 } 00:12:58.677 ] 00:12:58.677 }' 00:12:58.677 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:58.677 16:00:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:58.935 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:58.935 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:58.935 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:58.935 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:58.935 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:58.935 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:58.935 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:58.935 16:00:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.935 16:00:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:58.935 16:00:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.935 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:58.935 "name": "raid_bdev1", 00:12:58.935 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:58.935 "strip_size_kb": 0, 00:12:58.935 "state": "online", 00:12:58.935 "raid_level": "raid1", 00:12:58.935 "superblock": true, 00:12:58.935 "num_base_bdevs": 2, 00:12:58.935 "num_base_bdevs_discovered": 1, 00:12:58.935 "num_base_bdevs_operational": 1, 00:12:58.935 "base_bdevs_list": [ 00:12:58.935 { 00:12:58.935 "name": null, 00:12:58.935 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:58.935 "is_configured": false, 00:12:58.935 "data_offset": 0, 00:12:58.935 "data_size": 63488 00:12:58.935 }, 00:12:58.935 { 00:12:58.935 "name": "BaseBdev2", 00:12:58.935 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:58.935 "is_configured": true, 00:12:58.935 "data_offset": 2048, 00:12:58.935 "data_size": 63488 00:12:58.935 } 00:12:58.935 ] 00:12:58.935 }' 00:12:58.935 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:58.935 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:58.935 16:00:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:58.935 16:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:58.935 16:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:58.935 16:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:12:58.935 16:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:58.935 16:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:12:58.935 16:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:58.935 16:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:12:58.935 16:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:58.935 16:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:58.935 16:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.935 16:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:58.935 [2024-11-20 16:00:57.040529] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:58.935 [2024-11-20 16:00:57.040662] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:12:58.935 [2024-11-20 16:00:57.040692] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:58.935 request: 00:12:58.935 { 00:12:58.935 "base_bdev": "BaseBdev1", 00:12:58.935 "raid_bdev": "raid_bdev1", 00:12:58.935 "method": "bdev_raid_add_base_bdev", 00:12:58.935 "req_id": 1 00:12:58.935 } 00:12:58.935 Got JSON-RPC error response 00:12:58.935 response: 00:12:58.935 { 00:12:58.935 "code": -22, 00:12:58.935 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:12:58.935 } 00:12:58.935 16:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:12:58.935 16:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:12:58.935 16:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:12:58.935 16:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:12:58.935 16:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:12:58.935 16:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:12:59.867 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:59.867 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:59.867 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:59.867 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:59.867 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:59.867 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:59.867 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:59.867 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:59.867 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:59.867 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:59.867 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:59.867 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:59.867 16:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.867 16:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:59.867 16:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.867 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:59.867 "name": "raid_bdev1", 00:12:59.867 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:12:59.867 "strip_size_kb": 0, 00:12:59.867 "state": "online", 00:12:59.867 "raid_level": "raid1", 00:12:59.867 "superblock": true, 00:12:59.867 "num_base_bdevs": 2, 00:12:59.867 "num_base_bdevs_discovered": 1, 00:12:59.867 "num_base_bdevs_operational": 1, 00:12:59.868 "base_bdevs_list": [ 00:12:59.868 { 00:12:59.868 "name": null, 00:12:59.868 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:59.868 "is_configured": false, 00:12:59.868 "data_offset": 0, 00:12:59.868 "data_size": 63488 00:12:59.868 }, 00:12:59.868 { 00:12:59.868 "name": "BaseBdev2", 00:12:59.868 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:12:59.868 "is_configured": true, 00:12:59.868 "data_offset": 2048, 00:12:59.868 "data_size": 63488 00:12:59.868 } 00:12:59.868 ] 00:12:59.868 }' 00:12:59.868 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:59.868 16:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:00.433 "name": "raid_bdev1", 00:13:00.433 "uuid": "0184731f-813b-40a9-9661-63b66258e7ce", 00:13:00.433 "strip_size_kb": 0, 00:13:00.433 "state": "online", 00:13:00.433 "raid_level": "raid1", 00:13:00.433 "superblock": true, 00:13:00.433 "num_base_bdevs": 2, 00:13:00.433 "num_base_bdevs_discovered": 1, 00:13:00.433 "num_base_bdevs_operational": 1, 00:13:00.433 "base_bdevs_list": [ 00:13:00.433 { 00:13:00.433 "name": null, 00:13:00.433 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:00.433 "is_configured": false, 00:13:00.433 "data_offset": 0, 00:13:00.433 "data_size": 63488 00:13:00.433 }, 00:13:00.433 { 00:13:00.433 "name": "BaseBdev2", 00:13:00.433 "uuid": "49344dfa-17c1-5ca7-9205-8ced23f9713a", 00:13:00.433 "is_configured": true, 00:13:00.433 "data_offset": 2048, 00:13:00.433 "data_size": 63488 00:13:00.433 } 00:13:00.433 ] 00:13:00.433 }' 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 73622 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 73622 ']' 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 73622 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73622 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:00.433 killing process with pid 73622 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73622' 00:13:00.433 Received shutdown signal, test time was about 60.000000 seconds 00:13:00.433 00:13:00.433 Latency(us) 00:13:00.433 [2024-11-20T16:00:58.684Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:00.433 [2024-11-20T16:00:58.684Z] =================================================================================================================== 00:13:00.433 [2024-11-20T16:00:58.684Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 73622 00:13:00.433 [2024-11-20 16:00:58.496912] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:00.433 16:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 73622 00:13:00.433 [2024-11-20 16:00:58.497022] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:00.433 [2024-11-20 16:00:58.497065] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:00.434 [2024-11-20 16:00:58.497076] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:13:00.434 [2024-11-20 16:00:58.646270] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:01.001 16:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:13:01.001 00:13:01.001 real 0m23.209s 00:13:01.001 user 0m26.129s 00:13:01.001 sys 0m3.775s 00:13:01.001 16:00:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:01.001 16:00:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:01.001 ************************************ 00:13:01.001 END TEST raid_rebuild_test_sb 00:13:01.001 ************************************ 00:13:01.267 16:00:59 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 2 false true true 00:13:01.267 16:00:59 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:13:01.267 16:00:59 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:01.267 16:00:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:01.267 ************************************ 00:13:01.267 START TEST raid_rebuild_test_io 00:13:01.267 ************************************ 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 false true true 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=74353 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 74353 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # '[' -z 74353 ']' 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:01.267 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:01.267 16:00:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.267 [2024-11-20 16:00:59.332075] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:13:01.267 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:01.267 Zero copy mechanism will not be used. 00:13:01.267 [2024-11-20 16:00:59.332189] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74353 ] 00:13:01.267 [2024-11-20 16:00:59.482745] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:01.525 [2024-11-20 16:00:59.570249] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:01.526 [2024-11-20 16:00:59.683134] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:01.526 [2024-11-20 16:00:59.683168] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # return 0 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.093 BaseBdev1_malloc 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.093 [2024-11-20 16:01:00.250387] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:02.093 [2024-11-20 16:01:00.250442] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:02.093 [2024-11-20 16:01:00.250460] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:02.093 [2024-11-20 16:01:00.250469] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:02.093 [2024-11-20 16:01:00.252278] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:02.093 [2024-11-20 16:01:00.252312] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:02.093 BaseBdev1 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.093 BaseBdev2_malloc 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.093 [2024-11-20 16:01:00.286262] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:02.093 [2024-11-20 16:01:00.286317] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:02.093 [2024-11-20 16:01:00.286335] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:02.093 [2024-11-20 16:01:00.286344] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:02.093 [2024-11-20 16:01:00.288146] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:02.093 [2024-11-20 16:01:00.288178] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:02.093 BaseBdev2 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.093 spare_malloc 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.093 spare_delay 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.093 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.093 [2024-11-20 16:01:00.338357] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:02.093 [2024-11-20 16:01:00.338413] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:02.093 [2024-11-20 16:01:00.338429] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:13:02.093 [2024-11-20 16:01:00.338438] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:02.093 [2024-11-20 16:01:00.340258] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:02.093 [2024-11-20 16:01:00.340293] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:02.351 spare 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.351 [2024-11-20 16:01:00.346401] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:02.351 [2024-11-20 16:01:00.347943] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:02.351 [2024-11-20 16:01:00.348023] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:13:02.351 [2024-11-20 16:01:00.348034] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:13:02.351 [2024-11-20 16:01:00.348254] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:13:02.351 [2024-11-20 16:01:00.348383] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:13:02.351 [2024-11-20 16:01:00.348398] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:13:02.351 [2024-11-20 16:01:00.348524] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:02.351 "name": "raid_bdev1", 00:13:02.351 "uuid": "20e6ad44-616e-4433-89cf-a48a37696760", 00:13:02.351 "strip_size_kb": 0, 00:13:02.351 "state": "online", 00:13:02.351 "raid_level": "raid1", 00:13:02.351 "superblock": false, 00:13:02.351 "num_base_bdevs": 2, 00:13:02.351 "num_base_bdevs_discovered": 2, 00:13:02.351 "num_base_bdevs_operational": 2, 00:13:02.351 "base_bdevs_list": [ 00:13:02.351 { 00:13:02.351 "name": "BaseBdev1", 00:13:02.351 "uuid": "af530d47-b026-57b3-9ff0-cc5721d4351e", 00:13:02.351 "is_configured": true, 00:13:02.351 "data_offset": 0, 00:13:02.351 "data_size": 65536 00:13:02.351 }, 00:13:02.351 { 00:13:02.351 "name": "BaseBdev2", 00:13:02.351 "uuid": "d1734fff-f7dc-52c2-88fe-8d520a895512", 00:13:02.351 "is_configured": true, 00:13:02.351 "data_offset": 0, 00:13:02.351 "data_size": 65536 00:13:02.351 } 00:13:02.351 ] 00:13:02.351 }' 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:02.351 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.609 [2024-11-20 16:01:00.666702] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.609 [2024-11-20 16:01:00.726454] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:02.609 "name": "raid_bdev1", 00:13:02.609 "uuid": "20e6ad44-616e-4433-89cf-a48a37696760", 00:13:02.609 "strip_size_kb": 0, 00:13:02.609 "state": "online", 00:13:02.609 "raid_level": "raid1", 00:13:02.609 "superblock": false, 00:13:02.609 "num_base_bdevs": 2, 00:13:02.609 "num_base_bdevs_discovered": 1, 00:13:02.609 "num_base_bdevs_operational": 1, 00:13:02.609 "base_bdevs_list": [ 00:13:02.609 { 00:13:02.609 "name": null, 00:13:02.609 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.609 "is_configured": false, 00:13:02.609 "data_offset": 0, 00:13:02.609 "data_size": 65536 00:13:02.609 }, 00:13:02.609 { 00:13:02.609 "name": "BaseBdev2", 00:13:02.609 "uuid": "d1734fff-f7dc-52c2-88fe-8d520a895512", 00:13:02.609 "is_configured": true, 00:13:02.609 "data_offset": 0, 00:13:02.609 "data_size": 65536 00:13:02.609 } 00:13:02.609 ] 00:13:02.609 }' 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:02.609 16:01:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.609 [2024-11-20 16:01:00.819060] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:13:02.609 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:02.609 Zero copy mechanism will not be used. 00:13:02.609 Running I/O for 60 seconds... 00:13:02.867 16:01:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:02.867 16:01:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.867 16:01:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.867 [2024-11-20 16:01:01.054596] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:02.867 16:01:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.867 16:01:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:02.867 [2024-11-20 16:01:01.104015] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:13:02.867 [2024-11-20 16:01:01.105634] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:03.125 [2024-11-20 16:01:01.212841] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:03.125 [2024-11-20 16:01:01.213240] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:03.382 [2024-11-20 16:01:01.421507] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:03.382 [2024-11-20 16:01:01.421760] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:03.639 [2024-11-20 16:01:01.763208] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:03.897 187.00 IOPS, 561.00 MiB/s [2024-11-20T16:01:02.148Z] [2024-11-20 16:01:01.988331] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:03.897 [2024-11-20 16:01:01.988763] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:03.897 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:03.897 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:03.897 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:03.897 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:03.897 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:03.897 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:03.897 16:01:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.897 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:03.897 16:01:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.897 16:01:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.897 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:03.897 "name": "raid_bdev1", 00:13:03.897 "uuid": "20e6ad44-616e-4433-89cf-a48a37696760", 00:13:03.897 "strip_size_kb": 0, 00:13:03.897 "state": "online", 00:13:03.897 "raid_level": "raid1", 00:13:03.897 "superblock": false, 00:13:03.897 "num_base_bdevs": 2, 00:13:03.897 "num_base_bdevs_discovered": 2, 00:13:03.897 "num_base_bdevs_operational": 2, 00:13:03.897 "process": { 00:13:03.897 "type": "rebuild", 00:13:03.897 "target": "spare", 00:13:03.897 "progress": { 00:13:03.897 "blocks": 14336, 00:13:03.897 "percent": 21 00:13:03.897 } 00:13:03.897 }, 00:13:03.897 "base_bdevs_list": [ 00:13:03.897 { 00:13:03.897 "name": "spare", 00:13:03.897 "uuid": "c24b26fe-1d45-50b1-bc44-ba3e72f8cfb0", 00:13:03.897 "is_configured": true, 00:13:03.897 "data_offset": 0, 00:13:03.897 "data_size": 65536 00:13:03.897 }, 00:13:03.897 { 00:13:03.897 "name": "BaseBdev2", 00:13:03.897 "uuid": "d1734fff-f7dc-52c2-88fe-8d520a895512", 00:13:03.897 "is_configured": true, 00:13:03.897 "data_offset": 0, 00:13:03.897 "data_size": 65536 00:13:03.897 } 00:13:03.897 ] 00:13:03.897 }' 00:13:03.897 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.154 [2024-11-20 16:01:02.194655] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:04.154 [2024-11-20 16:01:02.308726] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:04.154 [2024-11-20 16:01:02.315897] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:04.154 [2024-11-20 16:01:02.315958] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:04.154 [2024-11-20 16:01:02.315970] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:04.154 [2024-11-20 16:01:02.348154] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006080 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:04.154 "name": "raid_bdev1", 00:13:04.154 "uuid": "20e6ad44-616e-4433-89cf-a48a37696760", 00:13:04.154 "strip_size_kb": 0, 00:13:04.154 "state": "online", 00:13:04.154 "raid_level": "raid1", 00:13:04.154 "superblock": false, 00:13:04.154 "num_base_bdevs": 2, 00:13:04.154 "num_base_bdevs_discovered": 1, 00:13:04.154 "num_base_bdevs_operational": 1, 00:13:04.154 "base_bdevs_list": [ 00:13:04.154 { 00:13:04.154 "name": null, 00:13:04.154 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:04.154 "is_configured": false, 00:13:04.154 "data_offset": 0, 00:13:04.154 "data_size": 65536 00:13:04.154 }, 00:13:04.154 { 00:13:04.154 "name": "BaseBdev2", 00:13:04.154 "uuid": "d1734fff-f7dc-52c2-88fe-8d520a895512", 00:13:04.154 "is_configured": true, 00:13:04.154 "data_offset": 0, 00:13:04.154 "data_size": 65536 00:13:04.154 } 00:13:04.154 ] 00:13:04.154 }' 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:04.154 16:01:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:04.719 "name": "raid_bdev1", 00:13:04.719 "uuid": "20e6ad44-616e-4433-89cf-a48a37696760", 00:13:04.719 "strip_size_kb": 0, 00:13:04.719 "state": "online", 00:13:04.719 "raid_level": "raid1", 00:13:04.719 "superblock": false, 00:13:04.719 "num_base_bdevs": 2, 00:13:04.719 "num_base_bdevs_discovered": 1, 00:13:04.719 "num_base_bdevs_operational": 1, 00:13:04.719 "base_bdevs_list": [ 00:13:04.719 { 00:13:04.719 "name": null, 00:13:04.719 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:04.719 "is_configured": false, 00:13:04.719 "data_offset": 0, 00:13:04.719 "data_size": 65536 00:13:04.719 }, 00:13:04.719 { 00:13:04.719 "name": "BaseBdev2", 00:13:04.719 "uuid": "d1734fff-f7dc-52c2-88fe-8d520a895512", 00:13:04.719 "is_configured": true, 00:13:04.719 "data_offset": 0, 00:13:04.719 "data_size": 65536 00:13:04.719 } 00:13:04.719 ] 00:13:04.719 }' 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.719 [2024-11-20 16:01:02.809919] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:04.719 16:01:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:04.719 188.00 IOPS, 564.00 MiB/s [2024-11-20T16:01:02.970Z] [2024-11-20 16:01:02.852169] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:13:04.719 [2024-11-20 16:01:02.853812] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:04.977 [2024-11-20 16:01:02.970850] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:04.977 [2024-11-20 16:01:02.971252] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:04.977 [2024-11-20 16:01:03.178005] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:04.977 [2024-11-20 16:01:03.178251] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:05.544 [2024-11-20 16:01:03.511965] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:05.544 [2024-11-20 16:01:03.636331] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:05.803 152.00 IOPS, 456.00 MiB/s [2024-11-20T16:01:04.054Z] 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:05.803 "name": "raid_bdev1", 00:13:05.803 "uuid": "20e6ad44-616e-4433-89cf-a48a37696760", 00:13:05.803 "strip_size_kb": 0, 00:13:05.803 "state": "online", 00:13:05.803 "raid_level": "raid1", 00:13:05.803 "superblock": false, 00:13:05.803 "num_base_bdevs": 2, 00:13:05.803 "num_base_bdevs_discovered": 2, 00:13:05.803 "num_base_bdevs_operational": 2, 00:13:05.803 "process": { 00:13:05.803 "type": "rebuild", 00:13:05.803 "target": "spare", 00:13:05.803 "progress": { 00:13:05.803 "blocks": 12288, 00:13:05.803 "percent": 18 00:13:05.803 } 00:13:05.803 }, 00:13:05.803 "base_bdevs_list": [ 00:13:05.803 { 00:13:05.803 "name": "spare", 00:13:05.803 "uuid": "c24b26fe-1d45-50b1-bc44-ba3e72f8cfb0", 00:13:05.803 "is_configured": true, 00:13:05.803 "data_offset": 0, 00:13:05.803 "data_size": 65536 00:13:05.803 }, 00:13:05.803 { 00:13:05.803 "name": "BaseBdev2", 00:13:05.803 "uuid": "d1734fff-f7dc-52c2-88fe-8d520a895512", 00:13:05.803 "is_configured": true, 00:13:05.803 "data_offset": 0, 00:13:05.803 "data_size": 65536 00:13:05.803 } 00:13:05.803 ] 00:13:05.803 }' 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:05.803 [2024-11-20 16:01:03.875371] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=317 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:05.803 "name": "raid_bdev1", 00:13:05.803 "uuid": "20e6ad44-616e-4433-89cf-a48a37696760", 00:13:05.803 "strip_size_kb": 0, 00:13:05.803 "state": "online", 00:13:05.803 "raid_level": "raid1", 00:13:05.803 "superblock": false, 00:13:05.803 "num_base_bdevs": 2, 00:13:05.803 "num_base_bdevs_discovered": 2, 00:13:05.803 "num_base_bdevs_operational": 2, 00:13:05.803 "process": { 00:13:05.803 "type": "rebuild", 00:13:05.803 "target": "spare", 00:13:05.803 "progress": { 00:13:05.803 "blocks": 14336, 00:13:05.803 "percent": 21 00:13:05.803 } 00:13:05.803 }, 00:13:05.803 "base_bdevs_list": [ 00:13:05.803 { 00:13:05.803 "name": "spare", 00:13:05.803 "uuid": "c24b26fe-1d45-50b1-bc44-ba3e72f8cfb0", 00:13:05.803 "is_configured": true, 00:13:05.803 "data_offset": 0, 00:13:05.803 "data_size": 65536 00:13:05.803 }, 00:13:05.803 { 00:13:05.803 "name": "BaseBdev2", 00:13:05.803 "uuid": "d1734fff-f7dc-52c2-88fe-8d520a895512", 00:13:05.803 "is_configured": true, 00:13:05.803 "data_offset": 0, 00:13:05.803 "data_size": 65536 00:13:05.803 } 00:13:05.803 ] 00:13:05.803 }' 00:13:05.803 16:01:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:05.803 16:01:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:05.803 16:01:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:06.061 16:01:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:06.061 16:01:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:06.061 [2024-11-20 16:01:04.096901] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:06.320 [2024-11-20 16:01:04.326954] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:13:06.320 [2024-11-20 16:01:04.327362] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:13:06.320 [2024-11-20 16:01:04.447182] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:13:06.835 133.25 IOPS, 399.75 MiB/s [2024-11-20T16:01:05.086Z] 16:01:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:06.835 16:01:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:06.835 16:01:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:06.835 16:01:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:06.835 16:01:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:06.835 16:01:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:07.094 16:01:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:07.094 16:01:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.094 16:01:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:07.094 16:01:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:07.094 16:01:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:07.094 16:01:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:07.094 "name": "raid_bdev1", 00:13:07.094 "uuid": "20e6ad44-616e-4433-89cf-a48a37696760", 00:13:07.094 "strip_size_kb": 0, 00:13:07.094 "state": "online", 00:13:07.094 "raid_level": "raid1", 00:13:07.094 "superblock": false, 00:13:07.094 "num_base_bdevs": 2, 00:13:07.094 "num_base_bdevs_discovered": 2, 00:13:07.094 "num_base_bdevs_operational": 2, 00:13:07.094 "process": { 00:13:07.094 "type": "rebuild", 00:13:07.094 "target": "spare", 00:13:07.094 "progress": { 00:13:07.094 "blocks": 30720, 00:13:07.094 "percent": 46 00:13:07.094 } 00:13:07.094 }, 00:13:07.094 "base_bdevs_list": [ 00:13:07.094 { 00:13:07.094 "name": "spare", 00:13:07.094 "uuid": "c24b26fe-1d45-50b1-bc44-ba3e72f8cfb0", 00:13:07.094 "is_configured": true, 00:13:07.094 "data_offset": 0, 00:13:07.094 "data_size": 65536 00:13:07.094 }, 00:13:07.094 { 00:13:07.094 "name": "BaseBdev2", 00:13:07.094 "uuid": "d1734fff-f7dc-52c2-88fe-8d520a895512", 00:13:07.094 "is_configured": true, 00:13:07.094 "data_offset": 0, 00:13:07.094 "data_size": 65536 00:13:07.094 } 00:13:07.094 ] 00:13:07.094 }' 00:13:07.094 16:01:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:07.094 [2024-11-20 16:01:05.136511] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:13:07.094 16:01:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:07.094 16:01:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:07.094 16:01:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:07.094 16:01:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:07.352 [2024-11-20 16:01:05.343581] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:13:07.352 [2024-11-20 16:01:05.343833] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:13:07.609 [2024-11-20 16:01:05.689077] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:13:08.175 113.40 IOPS, 340.20 MiB/s [2024-11-20T16:01:06.426Z] [2024-11-20 16:01:06.139429] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:13:08.175 16:01:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:08.175 16:01:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:08.175 16:01:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:08.175 16:01:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:08.175 16:01:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:08.175 16:01:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:08.175 16:01:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:08.175 16:01:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:08.175 16:01:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:08.175 16:01:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:08.175 16:01:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:08.175 16:01:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:08.175 "name": "raid_bdev1", 00:13:08.175 "uuid": "20e6ad44-616e-4433-89cf-a48a37696760", 00:13:08.175 "strip_size_kb": 0, 00:13:08.175 "state": "online", 00:13:08.175 "raid_level": "raid1", 00:13:08.175 "superblock": false, 00:13:08.175 "num_base_bdevs": 2, 00:13:08.175 "num_base_bdevs_discovered": 2, 00:13:08.175 "num_base_bdevs_operational": 2, 00:13:08.175 "process": { 00:13:08.175 "type": "rebuild", 00:13:08.175 "target": "spare", 00:13:08.175 "progress": { 00:13:08.175 "blocks": 45056, 00:13:08.175 "percent": 68 00:13:08.175 } 00:13:08.175 }, 00:13:08.175 "base_bdevs_list": [ 00:13:08.175 { 00:13:08.175 "name": "spare", 00:13:08.175 "uuid": "c24b26fe-1d45-50b1-bc44-ba3e72f8cfb0", 00:13:08.175 "is_configured": true, 00:13:08.175 "data_offset": 0, 00:13:08.175 "data_size": 65536 00:13:08.175 }, 00:13:08.175 { 00:13:08.175 "name": "BaseBdev2", 00:13:08.175 "uuid": "d1734fff-f7dc-52c2-88fe-8d520a895512", 00:13:08.175 "is_configured": true, 00:13:08.175 "data_offset": 0, 00:13:08.175 "data_size": 65536 00:13:08.175 } 00:13:08.175 ] 00:13:08.175 }' 00:13:08.175 16:01:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:08.175 16:01:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:08.176 16:01:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:08.176 16:01:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:08.176 16:01:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:08.176 [2024-11-20 16:01:06.354595] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:13:08.741 [2024-11-20 16:01:06.687562] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 51200 offset_begin: 49152 offset_end: 55296 00:13:08.741 101.00 IOPS, 303.00 MiB/s [2024-11-20T16:01:06.992Z] [2024-11-20 16:01:06.894498] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:13:08.741 [2024-11-20 16:01:06.894756] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:13:08.999 [2024-11-20 16:01:07.246784] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:13:09.258 16:01:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:09.258 16:01:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:09.258 16:01:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:09.258 16:01:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:09.258 16:01:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:09.258 16:01:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:09.258 16:01:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:09.258 16:01:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.258 16:01:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.258 16:01:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:09.258 16:01:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.258 16:01:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:09.258 "name": "raid_bdev1", 00:13:09.258 "uuid": "20e6ad44-616e-4433-89cf-a48a37696760", 00:13:09.258 "strip_size_kb": 0, 00:13:09.258 "state": "online", 00:13:09.258 "raid_level": "raid1", 00:13:09.258 "superblock": false, 00:13:09.258 "num_base_bdevs": 2, 00:13:09.258 "num_base_bdevs_discovered": 2, 00:13:09.258 "num_base_bdevs_operational": 2, 00:13:09.258 "process": { 00:13:09.258 "type": "rebuild", 00:13:09.258 "target": "spare", 00:13:09.258 "progress": { 00:13:09.258 "blocks": 59392, 00:13:09.258 "percent": 90 00:13:09.258 } 00:13:09.258 }, 00:13:09.258 "base_bdevs_list": [ 00:13:09.258 { 00:13:09.258 "name": "spare", 00:13:09.258 "uuid": "c24b26fe-1d45-50b1-bc44-ba3e72f8cfb0", 00:13:09.258 "is_configured": true, 00:13:09.258 "data_offset": 0, 00:13:09.258 "data_size": 65536 00:13:09.258 }, 00:13:09.258 { 00:13:09.258 "name": "BaseBdev2", 00:13:09.258 "uuid": "d1734fff-f7dc-52c2-88fe-8d520a895512", 00:13:09.258 "is_configured": true, 00:13:09.258 "data_offset": 0, 00:13:09.258 "data_size": 65536 00:13:09.258 } 00:13:09.258 ] 00:13:09.258 }' 00:13:09.258 16:01:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:09.258 16:01:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:09.258 16:01:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:09.258 16:01:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:09.258 16:01:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:09.516 [2024-11-20 16:01:07.694889] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:09.773 [2024-11-20 16:01:07.801310] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:09.773 [2024-11-20 16:01:07.804008] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:10.339 90.43 IOPS, 271.29 MiB/s [2024-11-20T16:01:08.590Z] 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:10.339 "name": "raid_bdev1", 00:13:10.339 "uuid": "20e6ad44-616e-4433-89cf-a48a37696760", 00:13:10.339 "strip_size_kb": 0, 00:13:10.339 "state": "online", 00:13:10.339 "raid_level": "raid1", 00:13:10.339 "superblock": false, 00:13:10.339 "num_base_bdevs": 2, 00:13:10.339 "num_base_bdevs_discovered": 2, 00:13:10.339 "num_base_bdevs_operational": 2, 00:13:10.339 "base_bdevs_list": [ 00:13:10.339 { 00:13:10.339 "name": "spare", 00:13:10.339 "uuid": "c24b26fe-1d45-50b1-bc44-ba3e72f8cfb0", 00:13:10.339 "is_configured": true, 00:13:10.339 "data_offset": 0, 00:13:10.339 "data_size": 65536 00:13:10.339 }, 00:13:10.339 { 00:13:10.339 "name": "BaseBdev2", 00:13:10.339 "uuid": "d1734fff-f7dc-52c2-88fe-8d520a895512", 00:13:10.339 "is_configured": true, 00:13:10.339 "data_offset": 0, 00:13:10.339 "data_size": 65536 00:13:10.339 } 00:13:10.339 ] 00:13:10.339 }' 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:10.339 "name": "raid_bdev1", 00:13:10.339 "uuid": "20e6ad44-616e-4433-89cf-a48a37696760", 00:13:10.339 "strip_size_kb": 0, 00:13:10.339 "state": "online", 00:13:10.339 "raid_level": "raid1", 00:13:10.339 "superblock": false, 00:13:10.339 "num_base_bdevs": 2, 00:13:10.339 "num_base_bdevs_discovered": 2, 00:13:10.339 "num_base_bdevs_operational": 2, 00:13:10.339 "base_bdevs_list": [ 00:13:10.339 { 00:13:10.339 "name": "spare", 00:13:10.339 "uuid": "c24b26fe-1d45-50b1-bc44-ba3e72f8cfb0", 00:13:10.339 "is_configured": true, 00:13:10.339 "data_offset": 0, 00:13:10.339 "data_size": 65536 00:13:10.339 }, 00:13:10.339 { 00:13:10.339 "name": "BaseBdev2", 00:13:10.339 "uuid": "d1734fff-f7dc-52c2-88fe-8d520a895512", 00:13:10.339 "is_configured": true, 00:13:10.339 "data_offset": 0, 00:13:10.339 "data_size": 65536 00:13:10.339 } 00:13:10.339 ] 00:13:10.339 }' 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:10.339 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:10.340 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:10.340 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:10.340 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:10.340 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:10.340 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:10.340 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:10.340 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:10.340 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:10.340 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:10.340 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:10.340 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:10.340 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:10.340 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:10.340 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:10.340 16:01:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:10.340 16:01:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:10.340 16:01:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:10.598 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:10.598 "name": "raid_bdev1", 00:13:10.598 "uuid": "20e6ad44-616e-4433-89cf-a48a37696760", 00:13:10.598 "strip_size_kb": 0, 00:13:10.598 "state": "online", 00:13:10.598 "raid_level": "raid1", 00:13:10.598 "superblock": false, 00:13:10.598 "num_base_bdevs": 2, 00:13:10.598 "num_base_bdevs_discovered": 2, 00:13:10.598 "num_base_bdevs_operational": 2, 00:13:10.598 "base_bdevs_list": [ 00:13:10.598 { 00:13:10.598 "name": "spare", 00:13:10.598 "uuid": "c24b26fe-1d45-50b1-bc44-ba3e72f8cfb0", 00:13:10.598 "is_configured": true, 00:13:10.598 "data_offset": 0, 00:13:10.598 "data_size": 65536 00:13:10.598 }, 00:13:10.598 { 00:13:10.598 "name": "BaseBdev2", 00:13:10.598 "uuid": "d1734fff-f7dc-52c2-88fe-8d520a895512", 00:13:10.598 "is_configured": true, 00:13:10.598 "data_offset": 0, 00:13:10.598 "data_size": 65536 00:13:10.598 } 00:13:10.598 ] 00:13:10.598 }' 00:13:10.598 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:10.598 16:01:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:10.856 83.38 IOPS, 250.12 MiB/s [2024-11-20T16:01:09.107Z] 16:01:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:10.856 16:01:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:10.856 16:01:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:10.856 [2024-11-20 16:01:08.892555] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:10.856 [2024-11-20 16:01:08.892591] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:10.856 00:13:10.856 Latency(us) 00:13:10.856 [2024-11-20T16:01:09.107Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:10.856 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:13:10.856 raid_bdev1 : 8.17 82.11 246.34 0.00 0.00 16530.16 310.35 113730.17 00:13:10.856 [2024-11-20T16:01:09.107Z] =================================================================================================================== 00:13:10.856 [2024-11-20T16:01:09.107Z] Total : 82.11 246.34 0.00 0.00 16530.16 310.35 113730.17 00:13:10.856 { 00:13:10.856 "results": [ 00:13:10.856 { 00:13:10.856 "job": "raid_bdev1", 00:13:10.856 "core_mask": "0x1", 00:13:10.856 "workload": "randrw", 00:13:10.856 "percentage": 50, 00:13:10.856 "status": "finished", 00:13:10.856 "queue_depth": 2, 00:13:10.856 "io_size": 3145728, 00:13:10.856 "runtime": 8.171673, 00:13:10.856 "iops": 82.11292840523599, 00:13:10.856 "mibps": 246.33878521570796, 00:13:10.856 "io_failed": 0, 00:13:10.857 "io_timeout": 0, 00:13:10.857 "avg_latency_us": 16530.158596813024, 00:13:10.857 "min_latency_us": 310.35076923076923, 00:13:10.857 "max_latency_us": 113730.16615384615 00:13:10.857 } 00:13:10.857 ], 00:13:10.857 "core_count": 1 00:13:10.857 } 00:13:10.857 [2024-11-20 16:01:09.008148] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:10.857 [2024-11-20 16:01:09.008213] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:10.857 [2024-11-20 16:01:09.008300] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:10.857 [2024-11-20 16:01:09.008310] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:10.857 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:13:11.116 /dev/nbd0 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:11.116 1+0 records in 00:13:11.116 1+0 records out 00:13:11.116 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000251937 s, 16.3 MB/s 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:11.116 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:13:11.377 /dev/nbd1 00:13:11.377 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:11.377 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:11.377 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:11.377 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:13:11.377 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:11.377 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:11.377 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:11.377 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:13:11.377 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:11.377 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:11.377 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:11.377 1+0 records in 00:13:11.377 1+0 records out 00:13:11.378 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000288651 s, 14.2 MB/s 00:13:11.378 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:11.378 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:13:11.378 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:11.378 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:11.378 16:01:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:13:11.378 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:11.378 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:11.378 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:11.637 16:01:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 74353 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # '[' -z 74353 ']' 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # kill -0 74353 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # uname 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74353 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:11.895 killing process with pid 74353 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74353' 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@973 -- # kill 74353 00:13:11.895 Received shutdown signal, test time was about 9.295733 seconds 00:13:11.895 00:13:11.895 Latency(us) 00:13:11.895 [2024-11-20T16:01:10.146Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:11.895 [2024-11-20T16:01:10.146Z] =================================================================================================================== 00:13:11.895 [2024-11-20T16:01:10.146Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:13:11.895 [2024-11-20 16:01:10.116558] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:11.895 16:01:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@978 -- # wait 74353 00:13:12.153 [2024-11-20 16:01:10.258012] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:13.085 16:01:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:13:13.085 00:13:13.085 real 0m11.736s 00:13:13.085 user 0m14.396s 00:13:13.085 sys 0m1.031s 00:13:13.085 16:01:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:13.085 16:01:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.085 ************************************ 00:13:13.085 END TEST raid_rebuild_test_io 00:13:13.085 ************************************ 00:13:13.085 16:01:11 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 2 true true true 00:13:13.085 16:01:11 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:13:13.085 16:01:11 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:13.085 16:01:11 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:13.085 ************************************ 00:13:13.085 START TEST raid_rebuild_test_sb_io 00:13:13.085 ************************************ 00:13:13.085 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true true true 00:13:13.085 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:13:13.085 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:13:13.085 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:13:13.085 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=74737 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 74737 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # '[' -z 74737 ']' 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:13.086 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.086 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:13.086 [2024-11-20 16:01:11.141892] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:13:13.086 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:13.086 Zero copy mechanism will not be used. 00:13:13.086 [2024-11-20 16:01:11.142072] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74737 ] 00:13:13.086 [2024-11-20 16:01:11.305397] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:13.344 [2024-11-20 16:01:11.406790] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:13.344 [2024-11-20 16:01:11.544756] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:13.344 [2024-11-20 16:01:11.544799] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:13.926 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:13.926 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # return 0 00:13:13.926 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:13.926 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:13.926 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.926 16:01:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.926 BaseBdev1_malloc 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.926 [2024-11-20 16:01:12.008271] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:13.926 [2024-11-20 16:01:12.008340] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:13.926 [2024-11-20 16:01:12.008363] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:13.926 [2024-11-20 16:01:12.008374] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:13.926 [2024-11-20 16:01:12.010555] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:13.926 [2024-11-20 16:01:12.010597] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:13.926 BaseBdev1 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.926 BaseBdev2_malloc 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.926 [2024-11-20 16:01:12.044368] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:13.926 [2024-11-20 16:01:12.044437] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:13.926 [2024-11-20 16:01:12.044461] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:13.926 [2024-11-20 16:01:12.044473] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:13.926 [2024-11-20 16:01:12.046624] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:13.926 [2024-11-20 16:01:12.046664] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:13.926 BaseBdev2 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.926 spare_malloc 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.926 spare_delay 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.926 [2024-11-20 16:01:12.105061] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:13.926 [2024-11-20 16:01:12.105130] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:13.926 [2024-11-20 16:01:12.105150] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:13:13.926 [2024-11-20 16:01:12.105162] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:13.926 [2024-11-20 16:01:12.107327] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:13.926 [2024-11-20 16:01:12.107368] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:13.926 spare 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.926 [2024-11-20 16:01:12.113127] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:13.926 [2024-11-20 16:01:12.114973] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:13.926 [2024-11-20 16:01:12.115148] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:13:13.926 [2024-11-20 16:01:12.115161] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:13.926 [2024-11-20 16:01:12.115429] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:13:13.926 [2024-11-20 16:01:12.115586] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:13:13.926 [2024-11-20 16:01:12.115595] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:13:13.926 [2024-11-20 16:01:12.115761] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.926 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:13.926 "name": "raid_bdev1", 00:13:13.926 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:13.926 "strip_size_kb": 0, 00:13:13.926 "state": "online", 00:13:13.926 "raid_level": "raid1", 00:13:13.926 "superblock": true, 00:13:13.926 "num_base_bdevs": 2, 00:13:13.926 "num_base_bdevs_discovered": 2, 00:13:13.926 "num_base_bdevs_operational": 2, 00:13:13.926 "base_bdevs_list": [ 00:13:13.926 { 00:13:13.926 "name": "BaseBdev1", 00:13:13.926 "uuid": "975c6765-323f-52e3-9187-90346e2439d7", 00:13:13.926 "is_configured": true, 00:13:13.926 "data_offset": 2048, 00:13:13.926 "data_size": 63488 00:13:13.926 }, 00:13:13.926 { 00:13:13.926 "name": "BaseBdev2", 00:13:13.926 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:13.926 "is_configured": true, 00:13:13.926 "data_offset": 2048, 00:13:13.927 "data_size": 63488 00:13:13.927 } 00:13:13.927 ] 00:13:13.927 }' 00:13:13.927 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:13.927 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:14.492 [2024-11-20 16:01:12.445452] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:14.492 [2024-11-20 16:01:12.505174] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:14.492 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.493 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:14.493 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.493 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:14.493 "name": "raid_bdev1", 00:13:14.493 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:14.493 "strip_size_kb": 0, 00:13:14.493 "state": "online", 00:13:14.493 "raid_level": "raid1", 00:13:14.493 "superblock": true, 00:13:14.493 "num_base_bdevs": 2, 00:13:14.493 "num_base_bdevs_discovered": 1, 00:13:14.493 "num_base_bdevs_operational": 1, 00:13:14.493 "base_bdevs_list": [ 00:13:14.493 { 00:13:14.493 "name": null, 00:13:14.493 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:14.493 "is_configured": false, 00:13:14.493 "data_offset": 0, 00:13:14.493 "data_size": 63488 00:13:14.493 }, 00:13:14.493 { 00:13:14.493 "name": "BaseBdev2", 00:13:14.493 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:14.493 "is_configured": true, 00:13:14.493 "data_offset": 2048, 00:13:14.493 "data_size": 63488 00:13:14.493 } 00:13:14.493 ] 00:13:14.493 }' 00:13:14.493 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:14.493 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:14.493 [2024-11-20 16:01:12.590568] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:13:14.493 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:14.493 Zero copy mechanism will not be used. 00:13:14.493 Running I/O for 60 seconds... 00:13:14.757 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:14.758 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.758 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:14.758 [2024-11-20 16:01:12.855492] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:14.758 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.758 16:01:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:14.758 [2024-11-20 16:01:12.912785] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:13:14.758 [2024-11-20 16:01:12.914728] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:15.015 [2024-11-20 16:01:13.016428] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:15.016 [2024-11-20 16:01:13.016887] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:15.016 [2024-11-20 16:01:13.139805] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:15.016 [2024-11-20 16:01:13.140060] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:15.276 [2024-11-20 16:01:13.370223] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:15.276 [2024-11-20 16:01:13.501732] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:15.276 [2024-11-20 16:01:13.501995] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:15.794 147.00 IOPS, 441.00 MiB/s [2024-11-20T16:01:14.045Z] 16:01:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:15.794 16:01:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:15.794 16:01:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:15.794 16:01:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:15.794 16:01:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:15.794 16:01:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.794 16:01:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.794 16:01:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.794 16:01:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:15.794 16:01:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.794 16:01:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:15.794 "name": "raid_bdev1", 00:13:15.794 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:15.794 "strip_size_kb": 0, 00:13:15.794 "state": "online", 00:13:15.794 "raid_level": "raid1", 00:13:15.794 "superblock": true, 00:13:15.794 "num_base_bdevs": 2, 00:13:15.794 "num_base_bdevs_discovered": 2, 00:13:15.794 "num_base_bdevs_operational": 2, 00:13:15.794 "process": { 00:13:15.794 "type": "rebuild", 00:13:15.794 "target": "spare", 00:13:15.794 "progress": { 00:13:15.794 "blocks": 14336, 00:13:15.794 "percent": 22 00:13:15.794 } 00:13:15.794 }, 00:13:15.794 "base_bdevs_list": [ 00:13:15.794 { 00:13:15.794 "name": "spare", 00:13:15.794 "uuid": "8f865218-aac6-5342-a037-8c64840892c5", 00:13:15.794 "is_configured": true, 00:13:15.794 "data_offset": 2048, 00:13:15.794 "data_size": 63488 00:13:15.794 }, 00:13:15.794 { 00:13:15.794 "name": "BaseBdev2", 00:13:15.794 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:15.794 "is_configured": true, 00:13:15.794 "data_offset": 2048, 00:13:15.794 "data_size": 63488 00:13:15.794 } 00:13:15.794 ] 00:13:15.794 }' 00:13:15.794 16:01:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:15.794 16:01:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:15.794 16:01:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:15.794 [2024-11-20 16:01:13.974663] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:15.794 [2024-11-20 16:01:13.974914] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:15.794 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:15.794 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:15.794 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.794 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.794 [2024-11-20 16:01:14.014900] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:16.053 [2024-11-20 16:01:14.115216] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:16.053 [2024-11-20 16:01:14.124078] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:16.053 [2024-11-20 16:01:14.124277] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:16.053 [2024-11-20 16:01:14.124310] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:16.053 [2024-11-20 16:01:14.157956] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006080 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:16.053 "name": "raid_bdev1", 00:13:16.053 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:16.053 "strip_size_kb": 0, 00:13:16.053 "state": "online", 00:13:16.053 "raid_level": "raid1", 00:13:16.053 "superblock": true, 00:13:16.053 "num_base_bdevs": 2, 00:13:16.053 "num_base_bdevs_discovered": 1, 00:13:16.053 "num_base_bdevs_operational": 1, 00:13:16.053 "base_bdevs_list": [ 00:13:16.053 { 00:13:16.053 "name": null, 00:13:16.053 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:16.053 "is_configured": false, 00:13:16.053 "data_offset": 0, 00:13:16.053 "data_size": 63488 00:13:16.053 }, 00:13:16.053 { 00:13:16.053 "name": "BaseBdev2", 00:13:16.053 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:16.053 "is_configured": true, 00:13:16.053 "data_offset": 2048, 00:13:16.053 "data_size": 63488 00:13:16.053 } 00:13:16.053 ] 00:13:16.053 }' 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:16.053 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:16.312 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:16.312 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:16.312 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:16.312 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:16.312 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:16.312 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:16.312 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.312 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.312 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:16.570 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.570 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:16.570 "name": "raid_bdev1", 00:13:16.570 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:16.570 "strip_size_kb": 0, 00:13:16.570 "state": "online", 00:13:16.570 "raid_level": "raid1", 00:13:16.570 "superblock": true, 00:13:16.570 "num_base_bdevs": 2, 00:13:16.570 "num_base_bdevs_discovered": 1, 00:13:16.570 "num_base_bdevs_operational": 1, 00:13:16.570 "base_bdevs_list": [ 00:13:16.570 { 00:13:16.570 "name": null, 00:13:16.570 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:16.570 "is_configured": false, 00:13:16.570 "data_offset": 0, 00:13:16.570 "data_size": 63488 00:13:16.570 }, 00:13:16.570 { 00:13:16.570 "name": "BaseBdev2", 00:13:16.570 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:16.570 "is_configured": true, 00:13:16.570 "data_offset": 2048, 00:13:16.570 "data_size": 63488 00:13:16.570 } 00:13:16.570 ] 00:13:16.570 }' 00:13:16.570 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:16.570 161.00 IOPS, 483.00 MiB/s [2024-11-20T16:01:14.821Z] 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:16.570 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:16.570 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:16.570 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:16.570 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.570 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:16.570 [2024-11-20 16:01:14.654506] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:16.570 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.570 16:01:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:16.570 [2024-11-20 16:01:14.698292] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:13:16.570 [2024-11-20 16:01:14.700226] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:16.829 [2024-11-20 16:01:14.822570] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:16.829 [2024-11-20 16:01:14.823034] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:16.829 [2024-11-20 16:01:15.032201] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:16.829 [2024-11-20 16:01:15.032472] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:17.394 [2024-11-20 16:01:15.373863] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:17.394 [2024-11-20 16:01:15.374300] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:17.394 153.00 IOPS, 459.00 MiB/s [2024-11-20T16:01:15.645Z] [2024-11-20 16:01:15.612620] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:17.652 "name": "raid_bdev1", 00:13:17.652 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:17.652 "strip_size_kb": 0, 00:13:17.652 "state": "online", 00:13:17.652 "raid_level": "raid1", 00:13:17.652 "superblock": true, 00:13:17.652 "num_base_bdevs": 2, 00:13:17.652 "num_base_bdevs_discovered": 2, 00:13:17.652 "num_base_bdevs_operational": 2, 00:13:17.652 "process": { 00:13:17.652 "type": "rebuild", 00:13:17.652 "target": "spare", 00:13:17.652 "progress": { 00:13:17.652 "blocks": 10240, 00:13:17.652 "percent": 16 00:13:17.652 } 00:13:17.652 }, 00:13:17.652 "base_bdevs_list": [ 00:13:17.652 { 00:13:17.652 "name": "spare", 00:13:17.652 "uuid": "8f865218-aac6-5342-a037-8c64840892c5", 00:13:17.652 "is_configured": true, 00:13:17.652 "data_offset": 2048, 00:13:17.652 "data_size": 63488 00:13:17.652 }, 00:13:17.652 { 00:13:17.652 "name": "BaseBdev2", 00:13:17.652 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:17.652 "is_configured": true, 00:13:17.652 "data_offset": 2048, 00:13:17.652 "data_size": 63488 00:13:17.652 } 00:13:17.652 ] 00:13:17.652 }' 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:13:17.652 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=329 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:17.652 "name": "raid_bdev1", 00:13:17.652 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:17.652 "strip_size_kb": 0, 00:13:17.652 "state": "online", 00:13:17.652 "raid_level": "raid1", 00:13:17.652 "superblock": true, 00:13:17.652 "num_base_bdevs": 2, 00:13:17.652 "num_base_bdevs_discovered": 2, 00:13:17.652 "num_base_bdevs_operational": 2, 00:13:17.652 "process": { 00:13:17.652 "type": "rebuild", 00:13:17.652 "target": "spare", 00:13:17.652 "progress": { 00:13:17.652 "blocks": 10240, 00:13:17.652 "percent": 16 00:13:17.652 } 00:13:17.652 }, 00:13:17.652 "base_bdevs_list": [ 00:13:17.652 { 00:13:17.652 "name": "spare", 00:13:17.652 "uuid": "8f865218-aac6-5342-a037-8c64840892c5", 00:13:17.652 "is_configured": true, 00:13:17.652 "data_offset": 2048, 00:13:17.652 "data_size": 63488 00:13:17.652 }, 00:13:17.652 { 00:13:17.652 "name": "BaseBdev2", 00:13:17.652 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:17.652 "is_configured": true, 00:13:17.652 "data_offset": 2048, 00:13:17.652 "data_size": 63488 00:13:17.652 } 00:13:17.652 ] 00:13:17.652 }' 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:17.652 16:01:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:17.910 [2024-11-20 16:01:15.947458] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:18.167 [2024-11-20 16:01:16.159535] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:18.167 [2024-11-20 16:01:16.159946] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:18.424 [2024-11-20 16:01:16.484849] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:13:18.424 [2024-11-20 16:01:16.598145] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:13:18.681 127.50 IOPS, 382.50 MiB/s [2024-11-20T16:01:16.932Z] [2024-11-20 16:01:16.823983] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:13:18.681 16:01:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:18.681 16:01:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:18.681 16:01:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:18.681 16:01:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:18.681 16:01:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:18.681 16:01:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:18.681 16:01:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:18.681 16:01:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:18.681 16:01:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:18.681 16:01:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.681 16:01:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:18.682 16:01:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:18.682 "name": "raid_bdev1", 00:13:18.682 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:18.682 "strip_size_kb": 0, 00:13:18.682 "state": "online", 00:13:18.682 "raid_level": "raid1", 00:13:18.682 "superblock": true, 00:13:18.682 "num_base_bdevs": 2, 00:13:18.682 "num_base_bdevs_discovered": 2, 00:13:18.682 "num_base_bdevs_operational": 2, 00:13:18.682 "process": { 00:13:18.682 "type": "rebuild", 00:13:18.682 "target": "spare", 00:13:18.682 "progress": { 00:13:18.682 "blocks": 26624, 00:13:18.682 "percent": 41 00:13:18.682 } 00:13:18.682 }, 00:13:18.682 "base_bdevs_list": [ 00:13:18.682 { 00:13:18.682 "name": "spare", 00:13:18.682 "uuid": "8f865218-aac6-5342-a037-8c64840892c5", 00:13:18.682 "is_configured": true, 00:13:18.682 "data_offset": 2048, 00:13:18.682 "data_size": 63488 00:13:18.682 }, 00:13:18.682 { 00:13:18.682 "name": "BaseBdev2", 00:13:18.682 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:18.682 "is_configured": true, 00:13:18.682 "data_offset": 2048, 00:13:18.682 "data_size": 63488 00:13:18.682 } 00:13:18.682 ] 00:13:18.682 }' 00:13:18.682 16:01:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:18.940 16:01:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:18.940 16:01:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:18.940 16:01:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:18.940 16:01:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:19.197 [2024-11-20 16:01:17.254851] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:13:19.197 [2024-11-20 16:01:17.255252] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:13:19.455 [2024-11-20 16:01:17.597953] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:13:19.714 113.00 IOPS, 339.00 MiB/s [2024-11-20T16:01:17.965Z] [2024-11-20 16:01:17.805487] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:13:19.714 [2024-11-20 16:01:17.805746] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:13:19.972 16:01:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:19.972 16:01:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:19.972 16:01:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:19.972 16:01:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:19.972 16:01:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:19.972 16:01:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:19.972 16:01:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:19.972 16:01:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:19.972 16:01:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.972 16:01:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.972 16:01:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.972 16:01:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:19.972 "name": "raid_bdev1", 00:13:19.972 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:19.972 "strip_size_kb": 0, 00:13:19.972 "state": "online", 00:13:19.972 "raid_level": "raid1", 00:13:19.972 "superblock": true, 00:13:19.972 "num_base_bdevs": 2, 00:13:19.972 "num_base_bdevs_discovered": 2, 00:13:19.972 "num_base_bdevs_operational": 2, 00:13:19.972 "process": { 00:13:19.972 "type": "rebuild", 00:13:19.972 "target": "spare", 00:13:19.972 "progress": { 00:13:19.972 "blocks": 43008, 00:13:19.972 "percent": 67 00:13:19.972 } 00:13:19.972 }, 00:13:19.972 "base_bdevs_list": [ 00:13:19.972 { 00:13:19.972 "name": "spare", 00:13:19.972 "uuid": "8f865218-aac6-5342-a037-8c64840892c5", 00:13:19.972 "is_configured": true, 00:13:19.972 "data_offset": 2048, 00:13:19.972 "data_size": 63488 00:13:19.972 }, 00:13:19.972 { 00:13:19.972 "name": "BaseBdev2", 00:13:19.972 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:19.972 "is_configured": true, 00:13:19.972 "data_offset": 2048, 00:13:19.972 "data_size": 63488 00:13:19.972 } 00:13:19.972 ] 00:13:19.972 }' 00:13:19.972 16:01:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:19.972 16:01:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:19.972 16:01:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:19.972 16:01:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:19.972 16:01:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:20.230 [2024-11-20 16:01:18.351741] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 51200 offset_begin: 49152 offset_end: 55296 00:13:20.747 99.67 IOPS, 299.00 MiB/s [2024-11-20T16:01:18.998Z] [2024-11-20 16:01:18.884002] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:13:21.005 16:01:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:21.005 16:01:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:21.005 16:01:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:21.005 16:01:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:21.005 16:01:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:21.005 16:01:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:21.005 16:01:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:21.005 16:01:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:21.005 16:01:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:21.005 16:01:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.005 16:01:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:21.005 [2024-11-20 16:01:19.103532] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:21.005 16:01:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:21.005 "name": "raid_bdev1", 00:13:21.005 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:21.005 "strip_size_kb": 0, 00:13:21.005 "state": "online", 00:13:21.005 "raid_level": "raid1", 00:13:21.005 "superblock": true, 00:13:21.005 "num_base_bdevs": 2, 00:13:21.005 "num_base_bdevs_discovered": 2, 00:13:21.005 "num_base_bdevs_operational": 2, 00:13:21.005 "process": { 00:13:21.005 "type": "rebuild", 00:13:21.005 "target": "spare", 00:13:21.005 "progress": { 00:13:21.005 "blocks": 61440, 00:13:21.005 "percent": 96 00:13:21.005 } 00:13:21.005 }, 00:13:21.005 "base_bdevs_list": [ 00:13:21.005 { 00:13:21.005 "name": "spare", 00:13:21.005 "uuid": "8f865218-aac6-5342-a037-8c64840892c5", 00:13:21.005 "is_configured": true, 00:13:21.005 "data_offset": 2048, 00:13:21.005 "data_size": 63488 00:13:21.005 }, 00:13:21.005 { 00:13:21.005 "name": "BaseBdev2", 00:13:21.005 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:21.005 "is_configured": true, 00:13:21.005 "data_offset": 2048, 00:13:21.005 "data_size": 63488 00:13:21.005 } 00:13:21.005 ] 00:13:21.005 }' 00:13:21.005 16:01:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:21.005 16:01:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:21.005 16:01:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:21.005 16:01:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:21.005 16:01:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:21.005 [2024-11-20 16:01:19.208595] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:21.005 [2024-11-20 16:01:19.210336] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:22.135 90.43 IOPS, 271.29 MiB/s [2024-11-20T16:01:20.386Z] 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:22.135 "name": "raid_bdev1", 00:13:22.135 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:22.135 "strip_size_kb": 0, 00:13:22.135 "state": "online", 00:13:22.135 "raid_level": "raid1", 00:13:22.135 "superblock": true, 00:13:22.135 "num_base_bdevs": 2, 00:13:22.135 "num_base_bdevs_discovered": 2, 00:13:22.135 "num_base_bdevs_operational": 2, 00:13:22.135 "base_bdevs_list": [ 00:13:22.135 { 00:13:22.135 "name": "spare", 00:13:22.135 "uuid": "8f865218-aac6-5342-a037-8c64840892c5", 00:13:22.135 "is_configured": true, 00:13:22.135 "data_offset": 2048, 00:13:22.135 "data_size": 63488 00:13:22.135 }, 00:13:22.135 { 00:13:22.135 "name": "BaseBdev2", 00:13:22.135 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:22.135 "is_configured": true, 00:13:22.135 "data_offset": 2048, 00:13:22.135 "data_size": 63488 00:13:22.135 } 00:13:22.135 ] 00:13:22.135 }' 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.135 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:22.136 "name": "raid_bdev1", 00:13:22.136 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:22.136 "strip_size_kb": 0, 00:13:22.136 "state": "online", 00:13:22.136 "raid_level": "raid1", 00:13:22.136 "superblock": true, 00:13:22.136 "num_base_bdevs": 2, 00:13:22.136 "num_base_bdevs_discovered": 2, 00:13:22.136 "num_base_bdevs_operational": 2, 00:13:22.136 "base_bdevs_list": [ 00:13:22.136 { 00:13:22.136 "name": "spare", 00:13:22.136 "uuid": "8f865218-aac6-5342-a037-8c64840892c5", 00:13:22.136 "is_configured": true, 00:13:22.136 "data_offset": 2048, 00:13:22.136 "data_size": 63488 00:13:22.136 }, 00:13:22.136 { 00:13:22.136 "name": "BaseBdev2", 00:13:22.136 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:22.136 "is_configured": true, 00:13:22.136 "data_offset": 2048, 00:13:22.136 "data_size": 63488 00:13:22.136 } 00:13:22.136 ] 00:13:22.136 }' 00:13:22.136 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:22.136 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:22.136 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:22.394 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:22.394 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:22.394 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:22.394 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:22.394 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:22.394 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:22.394 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:22.394 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:22.394 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:22.394 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:22.394 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:22.394 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.394 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:22.394 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.394 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:22.394 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.394 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:22.394 "name": "raid_bdev1", 00:13:22.394 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:22.394 "strip_size_kb": 0, 00:13:22.394 "state": "online", 00:13:22.394 "raid_level": "raid1", 00:13:22.394 "superblock": true, 00:13:22.395 "num_base_bdevs": 2, 00:13:22.395 "num_base_bdevs_discovered": 2, 00:13:22.395 "num_base_bdevs_operational": 2, 00:13:22.395 "base_bdevs_list": [ 00:13:22.395 { 00:13:22.395 "name": "spare", 00:13:22.395 "uuid": "8f865218-aac6-5342-a037-8c64840892c5", 00:13:22.395 "is_configured": true, 00:13:22.395 "data_offset": 2048, 00:13:22.395 "data_size": 63488 00:13:22.395 }, 00:13:22.395 { 00:13:22.395 "name": "BaseBdev2", 00:13:22.395 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:22.395 "is_configured": true, 00:13:22.395 "data_offset": 2048, 00:13:22.395 "data_size": 63488 00:13:22.395 } 00:13:22.395 ] 00:13:22.395 }' 00:13:22.395 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:22.395 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:22.653 83.88 IOPS, 251.62 MiB/s [2024-11-20T16:01:20.904Z] 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:22.653 [2024-11-20 16:01:20.771887] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:22.653 [2024-11-20 16:01:20.772005] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:22.653 00:13:22.653 Latency(us) 00:13:22.653 [2024-11-20T16:01:20.904Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:22.653 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:13:22.653 raid_bdev1 : 8.23 82.25 246.74 0.00 0.00 16565.61 332.41 116149.96 00:13:22.653 [2024-11-20T16:01:20.904Z] =================================================================================================================== 00:13:22.653 [2024-11-20T16:01:20.904Z] Total : 82.25 246.74 0.00 0.00 16565.61 332.41 116149.96 00:13:22.653 { 00:13:22.653 "results": [ 00:13:22.653 { 00:13:22.653 "job": "raid_bdev1", 00:13:22.653 "core_mask": "0x1", 00:13:22.653 "workload": "randrw", 00:13:22.653 "percentage": 50, 00:13:22.653 "status": "finished", 00:13:22.653 "queue_depth": 2, 00:13:22.653 "io_size": 3145728, 00:13:22.653 "runtime": 8.231246, 00:13:22.653 "iops": 82.24757223875947, 00:13:22.653 "mibps": 246.74271671627844, 00:13:22.653 "io_failed": 0, 00:13:22.653 "io_timeout": 0, 00:13:22.653 "avg_latency_us": 16565.60554482445, 00:13:22.653 "min_latency_us": 332.40615384615387, 00:13:22.653 "max_latency_us": 116149.95692307693 00:13:22.653 } 00:13:22.653 ], 00:13:22.653 "core_count": 1 00:13:22.653 } 00:13:22.653 [2024-11-20 16:01:20.835956] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:22.653 [2024-11-20 16:01:20.836015] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:22.653 [2024-11-20 16:01:20.836083] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:22.653 [2024-11-20 16:01:20.836093] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:22.653 16:01:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:13:22.911 /dev/nbd0 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:22.911 1+0 records in 00:13:22.911 1+0 records out 00:13:22.911 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000522816 s, 7.8 MB/s 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:22.911 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:13:23.169 /dev/nbd1 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:23.169 1+0 records in 00:13:23.169 1+0 records out 00:13:23.169 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00027162 s, 15.1 MB/s 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:23.169 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:23.427 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:23.427 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:23.427 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:23.427 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:23.427 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:23.427 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:23.427 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:23.427 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:23.427 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:23.427 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:23.427 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:23.427 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:23.427 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:23.427 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:23.427 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:23.427 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:23.427 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:23.428 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:23.428 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:23.428 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:23.428 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:23.428 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:23.685 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:23.685 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:23.685 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:23.685 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:23.685 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:23.685 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:23.685 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:23.686 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:23.686 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:13:23.686 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:13:23.686 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.686 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.686 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.686 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:23.686 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.686 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.686 [2024-11-20 16:01:21.882263] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:23.686 [2024-11-20 16:01:21.882463] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:23.686 [2024-11-20 16:01:21.882489] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:13:23.686 [2024-11-20 16:01:21.882499] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:23.686 [2024-11-20 16:01:21.884414] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:23.686 [2024-11-20 16:01:21.884454] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:23.686 [2024-11-20 16:01:21.884538] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:23.686 [2024-11-20 16:01:21.884584] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:23.686 [2024-11-20 16:01:21.884700] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:23.686 spare 00:13:23.686 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.686 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:13:23.686 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.686 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.943 [2024-11-20 16:01:21.984791] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:13:23.943 [2024-11-20 16:01:21.984966] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:23.943 [2024-11-20 16:01:21.985268] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b0d0 00:13:23.943 [2024-11-20 16:01:21.985472] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:13:23.943 [2024-11-20 16:01:21.985536] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:13:23.943 [2024-11-20 16:01:21.985739] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:23.943 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.943 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:23.943 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:23.943 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:23.943 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:23.943 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:23.943 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:23.943 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:23.943 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:23.943 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:23.943 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:23.943 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:23.943 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.943 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.943 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.943 16:01:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.943 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:23.943 "name": "raid_bdev1", 00:13:23.943 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:23.943 "strip_size_kb": 0, 00:13:23.943 "state": "online", 00:13:23.943 "raid_level": "raid1", 00:13:23.943 "superblock": true, 00:13:23.943 "num_base_bdevs": 2, 00:13:23.943 "num_base_bdevs_discovered": 2, 00:13:23.943 "num_base_bdevs_operational": 2, 00:13:23.943 "base_bdevs_list": [ 00:13:23.943 { 00:13:23.943 "name": "spare", 00:13:23.943 "uuid": "8f865218-aac6-5342-a037-8c64840892c5", 00:13:23.943 "is_configured": true, 00:13:23.943 "data_offset": 2048, 00:13:23.943 "data_size": 63488 00:13:23.943 }, 00:13:23.943 { 00:13:23.943 "name": "BaseBdev2", 00:13:23.943 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:23.943 "is_configured": true, 00:13:23.943 "data_offset": 2048, 00:13:23.943 "data_size": 63488 00:13:23.943 } 00:13:23.943 ] 00:13:23.943 }' 00:13:23.943 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:23.943 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:24.242 "name": "raid_bdev1", 00:13:24.242 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:24.242 "strip_size_kb": 0, 00:13:24.242 "state": "online", 00:13:24.242 "raid_level": "raid1", 00:13:24.242 "superblock": true, 00:13:24.242 "num_base_bdevs": 2, 00:13:24.242 "num_base_bdevs_discovered": 2, 00:13:24.242 "num_base_bdevs_operational": 2, 00:13:24.242 "base_bdevs_list": [ 00:13:24.242 { 00:13:24.242 "name": "spare", 00:13:24.242 "uuid": "8f865218-aac6-5342-a037-8c64840892c5", 00:13:24.242 "is_configured": true, 00:13:24.242 "data_offset": 2048, 00:13:24.242 "data_size": 63488 00:13:24.242 }, 00:13:24.242 { 00:13:24.242 "name": "BaseBdev2", 00:13:24.242 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:24.242 "is_configured": true, 00:13:24.242 "data_offset": 2048, 00:13:24.242 "data_size": 63488 00:13:24.242 } 00:13:24.242 ] 00:13:24.242 }' 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:24.242 [2024-11-20 16:01:22.422452] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:24.242 "name": "raid_bdev1", 00:13:24.242 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:24.242 "strip_size_kb": 0, 00:13:24.242 "state": "online", 00:13:24.242 "raid_level": "raid1", 00:13:24.242 "superblock": true, 00:13:24.242 "num_base_bdevs": 2, 00:13:24.242 "num_base_bdevs_discovered": 1, 00:13:24.242 "num_base_bdevs_operational": 1, 00:13:24.242 "base_bdevs_list": [ 00:13:24.242 { 00:13:24.242 "name": null, 00:13:24.242 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:24.242 "is_configured": false, 00:13:24.242 "data_offset": 0, 00:13:24.242 "data_size": 63488 00:13:24.242 }, 00:13:24.242 { 00:13:24.242 "name": "BaseBdev2", 00:13:24.242 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:24.242 "is_configured": true, 00:13:24.242 "data_offset": 2048, 00:13:24.242 "data_size": 63488 00:13:24.242 } 00:13:24.242 ] 00:13:24.242 }' 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:24.242 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:24.500 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:24.500 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.500 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:24.500 [2024-11-20 16:01:22.746566] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:24.500 [2024-11-20 16:01:22.746845] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:13:24.500 [2024-11-20 16:01:22.746863] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:24.500 [2024-11-20 16:01:22.746902] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:24.760 [2024-11-20 16:01:22.756086] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b1a0 00:13:24.760 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.760 16:01:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:13:24.760 [2024-11-20 16:01:22.757737] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:25.694 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:25.694 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:25.694 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:25.694 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:25.694 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:25.694 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:25.694 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:25.694 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.694 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:25.694 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.694 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:25.694 "name": "raid_bdev1", 00:13:25.694 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:25.694 "strip_size_kb": 0, 00:13:25.694 "state": "online", 00:13:25.694 "raid_level": "raid1", 00:13:25.694 "superblock": true, 00:13:25.694 "num_base_bdevs": 2, 00:13:25.694 "num_base_bdevs_discovered": 2, 00:13:25.694 "num_base_bdevs_operational": 2, 00:13:25.694 "process": { 00:13:25.694 "type": "rebuild", 00:13:25.694 "target": "spare", 00:13:25.694 "progress": { 00:13:25.694 "blocks": 20480, 00:13:25.694 "percent": 32 00:13:25.694 } 00:13:25.694 }, 00:13:25.694 "base_bdevs_list": [ 00:13:25.694 { 00:13:25.694 "name": "spare", 00:13:25.694 "uuid": "8f865218-aac6-5342-a037-8c64840892c5", 00:13:25.694 "is_configured": true, 00:13:25.694 "data_offset": 2048, 00:13:25.694 "data_size": 63488 00:13:25.694 }, 00:13:25.694 { 00:13:25.694 "name": "BaseBdev2", 00:13:25.694 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:25.694 "is_configured": true, 00:13:25.694 "data_offset": 2048, 00:13:25.694 "data_size": 63488 00:13:25.694 } 00:13:25.694 ] 00:13:25.694 }' 00:13:25.694 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:25.694 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:25.694 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:25.695 [2024-11-20 16:01:23.847882] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:25.695 [2024-11-20 16:01:23.863035] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:25.695 [2024-11-20 16:01:23.863221] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:25.695 [2024-11-20 16:01:23.863240] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:25.695 [2024-11-20 16:01:23.863247] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:25.695 "name": "raid_bdev1", 00:13:25.695 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:25.695 "strip_size_kb": 0, 00:13:25.695 "state": "online", 00:13:25.695 "raid_level": "raid1", 00:13:25.695 "superblock": true, 00:13:25.695 "num_base_bdevs": 2, 00:13:25.695 "num_base_bdevs_discovered": 1, 00:13:25.695 "num_base_bdevs_operational": 1, 00:13:25.695 "base_bdevs_list": [ 00:13:25.695 { 00:13:25.695 "name": null, 00:13:25.695 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:25.695 "is_configured": false, 00:13:25.695 "data_offset": 0, 00:13:25.695 "data_size": 63488 00:13:25.695 }, 00:13:25.695 { 00:13:25.695 "name": "BaseBdev2", 00:13:25.695 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:25.695 "is_configured": true, 00:13:25.695 "data_offset": 2048, 00:13:25.695 "data_size": 63488 00:13:25.695 } 00:13:25.695 ] 00:13:25.695 }' 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:25.695 16:01:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:25.953 16:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:25.953 16:01:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.953 16:01:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:25.953 [2024-11-20 16:01:24.191861] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:25.953 [2024-11-20 16:01:24.192023] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:25.953 [2024-11-20 16:01:24.192061] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:13:25.953 [2024-11-20 16:01:24.192111] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:25.953 [2024-11-20 16:01:24.192503] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:25.953 [2024-11-20 16:01:24.192518] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:25.953 [2024-11-20 16:01:24.192592] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:25.953 [2024-11-20 16:01:24.192603] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:13:25.953 [2024-11-20 16:01:24.192614] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:25.953 [2024-11-20 16:01:24.192633] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:26.210 spare 00:13:26.210 [2024-11-20 16:01:24.201800] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b270 00:13:26.210 16:01:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.210 16:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:13:26.210 [2024-11-20 16:01:24.203422] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:27.142 "name": "raid_bdev1", 00:13:27.142 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:27.142 "strip_size_kb": 0, 00:13:27.142 "state": "online", 00:13:27.142 "raid_level": "raid1", 00:13:27.142 "superblock": true, 00:13:27.142 "num_base_bdevs": 2, 00:13:27.142 "num_base_bdevs_discovered": 2, 00:13:27.142 "num_base_bdevs_operational": 2, 00:13:27.142 "process": { 00:13:27.142 "type": "rebuild", 00:13:27.142 "target": "spare", 00:13:27.142 "progress": { 00:13:27.142 "blocks": 20480, 00:13:27.142 "percent": 32 00:13:27.142 } 00:13:27.142 }, 00:13:27.142 "base_bdevs_list": [ 00:13:27.142 { 00:13:27.142 "name": "spare", 00:13:27.142 "uuid": "8f865218-aac6-5342-a037-8c64840892c5", 00:13:27.142 "is_configured": true, 00:13:27.142 "data_offset": 2048, 00:13:27.142 "data_size": 63488 00:13:27.142 }, 00:13:27.142 { 00:13:27.142 "name": "BaseBdev2", 00:13:27.142 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:27.142 "is_configured": true, 00:13:27.142 "data_offset": 2048, 00:13:27.142 "data_size": 63488 00:13:27.142 } 00:13:27.142 ] 00:13:27.142 }' 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.142 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.142 [2024-11-20 16:01:25.321897] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:27.399 [2024-11-20 16:01:25.409047] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:27.399 [2024-11-20 16:01:25.409256] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:27.399 [2024-11-20 16:01:25.409273] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:27.399 [2024-11-20 16:01:25.409282] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:27.399 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.399 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:27.399 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:27.399 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:27.399 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:27.399 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:27.399 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:13:27.399 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:27.399 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:27.399 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:27.399 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:27.400 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:27.400 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:27.400 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.400 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.400 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.400 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:27.400 "name": "raid_bdev1", 00:13:27.400 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:27.400 "strip_size_kb": 0, 00:13:27.400 "state": "online", 00:13:27.400 "raid_level": "raid1", 00:13:27.400 "superblock": true, 00:13:27.400 "num_base_bdevs": 2, 00:13:27.400 "num_base_bdevs_discovered": 1, 00:13:27.400 "num_base_bdevs_operational": 1, 00:13:27.400 "base_bdevs_list": [ 00:13:27.400 { 00:13:27.400 "name": null, 00:13:27.400 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:27.400 "is_configured": false, 00:13:27.400 "data_offset": 0, 00:13:27.400 "data_size": 63488 00:13:27.400 }, 00:13:27.400 { 00:13:27.400 "name": "BaseBdev2", 00:13:27.400 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:27.400 "is_configured": true, 00:13:27.400 "data_offset": 2048, 00:13:27.400 "data_size": 63488 00:13:27.400 } 00:13:27.400 ] 00:13:27.400 }' 00:13:27.400 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:27.400 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:27.657 "name": "raid_bdev1", 00:13:27.657 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:27.657 "strip_size_kb": 0, 00:13:27.657 "state": "online", 00:13:27.657 "raid_level": "raid1", 00:13:27.657 "superblock": true, 00:13:27.657 "num_base_bdevs": 2, 00:13:27.657 "num_base_bdevs_discovered": 1, 00:13:27.657 "num_base_bdevs_operational": 1, 00:13:27.657 "base_bdevs_list": [ 00:13:27.657 { 00:13:27.657 "name": null, 00:13:27.657 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:27.657 "is_configured": false, 00:13:27.657 "data_offset": 0, 00:13:27.657 "data_size": 63488 00:13:27.657 }, 00:13:27.657 { 00:13:27.657 "name": "BaseBdev2", 00:13:27.657 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:27.657 "is_configured": true, 00:13:27.657 "data_offset": 2048, 00:13:27.657 "data_size": 63488 00:13:27.657 } 00:13:27.657 ] 00:13:27.657 }' 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.657 [2024-11-20 16:01:25.849561] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:27.657 [2024-11-20 16:01:25.849609] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:27.657 [2024-11-20 16:01:25.849628] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:13:27.657 [2024-11-20 16:01:25.849637] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:27.657 [2024-11-20 16:01:25.849989] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:27.657 [2024-11-20 16:01:25.850007] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:27.657 [2024-11-20 16:01:25.850066] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:13:27.657 [2024-11-20 16:01:25.850078] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:13:27.657 [2024-11-20 16:01:25.850084] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:27.657 [2024-11-20 16:01:25.850095] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:13:27.657 BaseBdev1 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.657 16:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:29.030 "name": "raid_bdev1", 00:13:29.030 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:29.030 "strip_size_kb": 0, 00:13:29.030 "state": "online", 00:13:29.030 "raid_level": "raid1", 00:13:29.030 "superblock": true, 00:13:29.030 "num_base_bdevs": 2, 00:13:29.030 "num_base_bdevs_discovered": 1, 00:13:29.030 "num_base_bdevs_operational": 1, 00:13:29.030 "base_bdevs_list": [ 00:13:29.030 { 00:13:29.030 "name": null, 00:13:29.030 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:29.030 "is_configured": false, 00:13:29.030 "data_offset": 0, 00:13:29.030 "data_size": 63488 00:13:29.030 }, 00:13:29.030 { 00:13:29.030 "name": "BaseBdev2", 00:13:29.030 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:29.030 "is_configured": true, 00:13:29.030 "data_offset": 2048, 00:13:29.030 "data_size": 63488 00:13:29.030 } 00:13:29.030 ] 00:13:29.030 }' 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:29.030 16:01:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:29.030 "name": "raid_bdev1", 00:13:29.030 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:29.030 "strip_size_kb": 0, 00:13:29.030 "state": "online", 00:13:29.030 "raid_level": "raid1", 00:13:29.030 "superblock": true, 00:13:29.030 "num_base_bdevs": 2, 00:13:29.030 "num_base_bdevs_discovered": 1, 00:13:29.030 "num_base_bdevs_operational": 1, 00:13:29.030 "base_bdevs_list": [ 00:13:29.030 { 00:13:29.030 "name": null, 00:13:29.030 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:29.030 "is_configured": false, 00:13:29.030 "data_offset": 0, 00:13:29.030 "data_size": 63488 00:13:29.030 }, 00:13:29.030 { 00:13:29.030 "name": "BaseBdev2", 00:13:29.030 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:29.030 "is_configured": true, 00:13:29.030 "data_offset": 2048, 00:13:29.030 "data_size": 63488 00:13:29.030 } 00:13:29.030 ] 00:13:29.030 }' 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # local es=0 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:13:29.030 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:29.031 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:29.031 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.031 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:29.288 [2024-11-20 16:01:27.282052] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:29.288 [2024-11-20 16:01:27.282177] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:13:29.288 [2024-11-20 16:01:27.282190] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:29.288 request: 00:13:29.288 { 00:13:29.288 "base_bdev": "BaseBdev1", 00:13:29.288 "raid_bdev": "raid_bdev1", 00:13:29.288 "method": "bdev_raid_add_base_bdev", 00:13:29.288 "req_id": 1 00:13:29.288 } 00:13:29.288 Got JSON-RPC error response 00:13:29.288 response: 00:13:29.288 { 00:13:29.288 "code": -22, 00:13:29.288 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:13:29.288 } 00:13:29.288 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:13:29.288 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # es=1 00:13:29.288 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:13:29.288 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:13:29.288 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:13:29.288 16:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:30.222 "name": "raid_bdev1", 00:13:30.222 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:30.222 "strip_size_kb": 0, 00:13:30.222 "state": "online", 00:13:30.222 "raid_level": "raid1", 00:13:30.222 "superblock": true, 00:13:30.222 "num_base_bdevs": 2, 00:13:30.222 "num_base_bdevs_discovered": 1, 00:13:30.222 "num_base_bdevs_operational": 1, 00:13:30.222 "base_bdevs_list": [ 00:13:30.222 { 00:13:30.222 "name": null, 00:13:30.222 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:30.222 "is_configured": false, 00:13:30.222 "data_offset": 0, 00:13:30.222 "data_size": 63488 00:13:30.222 }, 00:13:30.222 { 00:13:30.222 "name": "BaseBdev2", 00:13:30.222 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:30.222 "is_configured": true, 00:13:30.222 "data_offset": 2048, 00:13:30.222 "data_size": 63488 00:13:30.222 } 00:13:30.222 ] 00:13:30.222 }' 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:30.222 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:30.481 "name": "raid_bdev1", 00:13:30.481 "uuid": "59852174-a8a3-4223-83f4-0bef2a40fa21", 00:13:30.481 "strip_size_kb": 0, 00:13:30.481 "state": "online", 00:13:30.481 "raid_level": "raid1", 00:13:30.481 "superblock": true, 00:13:30.481 "num_base_bdevs": 2, 00:13:30.481 "num_base_bdevs_discovered": 1, 00:13:30.481 "num_base_bdevs_operational": 1, 00:13:30.481 "base_bdevs_list": [ 00:13:30.481 { 00:13:30.481 "name": null, 00:13:30.481 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:30.481 "is_configured": false, 00:13:30.481 "data_offset": 0, 00:13:30.481 "data_size": 63488 00:13:30.481 }, 00:13:30.481 { 00:13:30.481 "name": "BaseBdev2", 00:13:30.481 "uuid": "df094b29-4baf-5c64-9ee1-2e78234ecfb3", 00:13:30.481 "is_configured": true, 00:13:30.481 "data_offset": 2048, 00:13:30.481 "data_size": 63488 00:13:30.481 } 00:13:30.481 ] 00:13:30.481 }' 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 74737 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # '[' -z 74737 ']' 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # kill -0 74737 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # uname 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74737 00:13:30.481 killing process with pid 74737 00:13:30.481 Received shutdown signal, test time was about 16.111452 seconds 00:13:30.481 00:13:30.481 Latency(us) 00:13:30.481 [2024-11-20T16:01:28.732Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:30.481 [2024-11-20T16:01:28.732Z] =================================================================================================================== 00:13:30.481 [2024-11-20T16:01:28.732Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74737' 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@973 -- # kill 74737 00:13:30.481 [2024-11-20 16:01:28.704083] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:30.481 16:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@978 -- # wait 74737 00:13:30.481 [2024-11-20 16:01:28.704185] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:30.481 [2024-11-20 16:01:28.704232] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:30.481 [2024-11-20 16:01:28.704239] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:13:30.738 [2024-11-20 16:01:28.815170] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:31.304 ************************************ 00:13:31.304 END TEST raid_rebuild_test_sb_io 00:13:31.304 ************************************ 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:13:31.304 00:13:31.304 real 0m18.368s 00:13:31.304 user 0m23.240s 00:13:31.304 sys 0m1.480s 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:31.304 16:01:29 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:13:31.304 16:01:29 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 4 false false true 00:13:31.304 16:01:29 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:13:31.304 16:01:29 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:31.304 16:01:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:31.304 ************************************ 00:13:31.304 START TEST raid_rebuild_test 00:13:31.304 ************************************ 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 false false true 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=75417 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 75417 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 75417 ']' 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:31.304 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:31.304 16:01:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.304 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:31.304 Zero copy mechanism will not be used. 00:13:31.304 [2024-11-20 16:01:29.530324] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:13:31.304 [2024-11-20 16:01:29.530445] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid75417 ] 00:13:31.629 [2024-11-20 16:01:29.693157] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:31.629 [2024-11-20 16:01:29.794121] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:31.890 [2024-11-20 16:01:29.929654] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:31.890 [2024-11-20 16:01:29.929708] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:32.149 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:32.149 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:13:32.149 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:32.149 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:32.149 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.149 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.409 BaseBdev1_malloc 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.409 [2024-11-20 16:01:30.431551] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:32.409 [2024-11-20 16:01:30.431724] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:32.409 [2024-11-20 16:01:30.431768] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:32.409 [2024-11-20 16:01:30.431829] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:32.409 [2024-11-20 16:01:30.434040] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:32.409 [2024-11-20 16:01:30.434142] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:32.409 BaseBdev1 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.409 BaseBdev2_malloc 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.409 [2024-11-20 16:01:30.475313] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:32.409 [2024-11-20 16:01:30.475440] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:32.409 [2024-11-20 16:01:30.475481] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:32.409 [2024-11-20 16:01:30.475534] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:32.409 [2024-11-20 16:01:30.477629] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:32.409 [2024-11-20 16:01:30.477741] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:32.409 BaseBdev2 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.409 BaseBdev3_malloc 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.409 [2024-11-20 16:01:30.523304] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:32.409 [2024-11-20 16:01:30.523355] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:32.409 [2024-11-20 16:01:30.523375] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:32.409 [2024-11-20 16:01:30.523386] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:32.409 [2024-11-20 16:01:30.525455] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:32.409 [2024-11-20 16:01:30.525492] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:32.409 BaseBdev3 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.409 BaseBdev4_malloc 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.409 [2024-11-20 16:01:30.563372] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:13:32.409 [2024-11-20 16:01:30.563424] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:32.409 [2024-11-20 16:01:30.563440] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:13:32.409 [2024-11-20 16:01:30.563450] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:32.409 [2024-11-20 16:01:30.565496] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:32.409 [2024-11-20 16:01:30.565533] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:13:32.409 BaseBdev4 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.409 spare_malloc 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.409 spare_delay 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.409 [2024-11-20 16:01:30.615248] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:32.409 [2024-11-20 16:01:30.615296] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:32.409 [2024-11-20 16:01:30.615313] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:13:32.409 [2024-11-20 16:01:30.615323] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:32.409 [2024-11-20 16:01:30.617391] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:32.409 [2024-11-20 16:01:30.617426] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:32.409 spare 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.409 [2024-11-20 16:01:30.623295] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:32.409 [2024-11-20 16:01:30.625150] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:32.409 [2024-11-20 16:01:30.625209] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:32.409 [2024-11-20 16:01:30.625259] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:32.409 [2024-11-20 16:01:30.625337] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:13:32.409 [2024-11-20 16:01:30.625349] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:13:32.409 [2024-11-20 16:01:30.625596] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:13:32.409 [2024-11-20 16:01:30.625753] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:13:32.409 [2024-11-20 16:01:30.625767] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:13:32.409 [2024-11-20 16:01:30.625901] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:32.409 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:32.410 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:32.410 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:32.410 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:32.410 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:32.410 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:32.410 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:32.410 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:32.410 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.410 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:32.410 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.410 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.410 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.668 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:32.668 "name": "raid_bdev1", 00:13:32.668 "uuid": "abb7ffcb-4de4-476d-bf11-501d0c2a10b2", 00:13:32.668 "strip_size_kb": 0, 00:13:32.668 "state": "online", 00:13:32.668 "raid_level": "raid1", 00:13:32.668 "superblock": false, 00:13:32.668 "num_base_bdevs": 4, 00:13:32.668 "num_base_bdevs_discovered": 4, 00:13:32.668 "num_base_bdevs_operational": 4, 00:13:32.668 "base_bdevs_list": [ 00:13:32.668 { 00:13:32.668 "name": "BaseBdev1", 00:13:32.668 "uuid": "ced53f1d-7c28-5f90-8005-d55ac5d771fa", 00:13:32.668 "is_configured": true, 00:13:32.668 "data_offset": 0, 00:13:32.668 "data_size": 65536 00:13:32.668 }, 00:13:32.668 { 00:13:32.668 "name": "BaseBdev2", 00:13:32.668 "uuid": "154a7b63-1ad7-5526-8b8f-8904d1af1a44", 00:13:32.668 "is_configured": true, 00:13:32.668 "data_offset": 0, 00:13:32.668 "data_size": 65536 00:13:32.668 }, 00:13:32.668 { 00:13:32.668 "name": "BaseBdev3", 00:13:32.668 "uuid": "9a487f7d-0585-539a-98e3-1f90b2da8abd", 00:13:32.668 "is_configured": true, 00:13:32.668 "data_offset": 0, 00:13:32.668 "data_size": 65536 00:13:32.668 }, 00:13:32.668 { 00:13:32.668 "name": "BaseBdev4", 00:13:32.668 "uuid": "36cfb9f6-2877-5bae-90d5-98d75414362e", 00:13:32.668 "is_configured": true, 00:13:32.668 "data_offset": 0, 00:13:32.668 "data_size": 65536 00:13:32.668 } 00:13:32.668 ] 00:13:32.668 }' 00:13:32.668 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:32.668 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.926 [2024-11-20 16:01:30.931748] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:32.926 16:01:30 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:13:33.186 [2024-11-20 16:01:31.183472] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:13:33.186 /dev/nbd0 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:33.186 1+0 records in 00:13:33.186 1+0 records out 00:13:33.186 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000263275 s, 15.6 MB/s 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:13:33.186 16:01:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:13:43.174 65536+0 records in 00:13:43.174 65536+0 records out 00:13:43.174 33554432 bytes (34 MB, 32 MiB) copied, 8.43631 s, 4.0 MB/s 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:43.174 [2024-11-20 16:01:39.865183] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.174 [2024-11-20 16:01:39.897265] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.174 16:01:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:43.174 "name": "raid_bdev1", 00:13:43.174 "uuid": "abb7ffcb-4de4-476d-bf11-501d0c2a10b2", 00:13:43.174 "strip_size_kb": 0, 00:13:43.174 "state": "online", 00:13:43.174 "raid_level": "raid1", 00:13:43.174 "superblock": false, 00:13:43.174 "num_base_bdevs": 4, 00:13:43.174 "num_base_bdevs_discovered": 3, 00:13:43.174 "num_base_bdevs_operational": 3, 00:13:43.175 "base_bdevs_list": [ 00:13:43.175 { 00:13:43.175 "name": null, 00:13:43.175 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:43.175 "is_configured": false, 00:13:43.175 "data_offset": 0, 00:13:43.175 "data_size": 65536 00:13:43.175 }, 00:13:43.175 { 00:13:43.175 "name": "BaseBdev2", 00:13:43.175 "uuid": "154a7b63-1ad7-5526-8b8f-8904d1af1a44", 00:13:43.175 "is_configured": true, 00:13:43.175 "data_offset": 0, 00:13:43.175 "data_size": 65536 00:13:43.175 }, 00:13:43.175 { 00:13:43.175 "name": "BaseBdev3", 00:13:43.175 "uuid": "9a487f7d-0585-539a-98e3-1f90b2da8abd", 00:13:43.175 "is_configured": true, 00:13:43.175 "data_offset": 0, 00:13:43.175 "data_size": 65536 00:13:43.175 }, 00:13:43.175 { 00:13:43.175 "name": "BaseBdev4", 00:13:43.175 "uuid": "36cfb9f6-2877-5bae-90d5-98d75414362e", 00:13:43.175 "is_configured": true, 00:13:43.175 "data_offset": 0, 00:13:43.175 "data_size": 65536 00:13:43.175 } 00:13:43.175 ] 00:13:43.175 }' 00:13:43.175 16:01:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:43.175 16:01:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.175 16:01:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:43.175 16:01:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.175 16:01:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.175 [2024-11-20 16:01:40.225353] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:43.175 [2024-11-20 16:01:40.235627] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09d70 00:13:43.175 16:01:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.175 16:01:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:43.175 [2024-11-20 16:01:40.237731] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:43.175 "name": "raid_bdev1", 00:13:43.175 "uuid": "abb7ffcb-4de4-476d-bf11-501d0c2a10b2", 00:13:43.175 "strip_size_kb": 0, 00:13:43.175 "state": "online", 00:13:43.175 "raid_level": "raid1", 00:13:43.175 "superblock": false, 00:13:43.175 "num_base_bdevs": 4, 00:13:43.175 "num_base_bdevs_discovered": 4, 00:13:43.175 "num_base_bdevs_operational": 4, 00:13:43.175 "process": { 00:13:43.175 "type": "rebuild", 00:13:43.175 "target": "spare", 00:13:43.175 "progress": { 00:13:43.175 "blocks": 20480, 00:13:43.175 "percent": 31 00:13:43.175 } 00:13:43.175 }, 00:13:43.175 "base_bdevs_list": [ 00:13:43.175 { 00:13:43.175 "name": "spare", 00:13:43.175 "uuid": "ea39c658-7bc1-5a69-b45e-bffc31a97f83", 00:13:43.175 "is_configured": true, 00:13:43.175 "data_offset": 0, 00:13:43.175 "data_size": 65536 00:13:43.175 }, 00:13:43.175 { 00:13:43.175 "name": "BaseBdev2", 00:13:43.175 "uuid": "154a7b63-1ad7-5526-8b8f-8904d1af1a44", 00:13:43.175 "is_configured": true, 00:13:43.175 "data_offset": 0, 00:13:43.175 "data_size": 65536 00:13:43.175 }, 00:13:43.175 { 00:13:43.175 "name": "BaseBdev3", 00:13:43.175 "uuid": "9a487f7d-0585-539a-98e3-1f90b2da8abd", 00:13:43.175 "is_configured": true, 00:13:43.175 "data_offset": 0, 00:13:43.175 "data_size": 65536 00:13:43.175 }, 00:13:43.175 { 00:13:43.175 "name": "BaseBdev4", 00:13:43.175 "uuid": "36cfb9f6-2877-5bae-90d5-98d75414362e", 00:13:43.175 "is_configured": true, 00:13:43.175 "data_offset": 0, 00:13:43.175 "data_size": 65536 00:13:43.175 } 00:13:43.175 ] 00:13:43.175 }' 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.175 16:01:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.175 [2024-11-20 16:01:41.351556] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:43.432 [2024-11-20 16:01:41.443598] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:43.432 [2024-11-20 16:01:41.443818] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:43.432 [2024-11-20 16:01:41.443836] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:43.432 [2024-11-20 16:01:41.443846] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:43.432 "name": "raid_bdev1", 00:13:43.432 "uuid": "abb7ffcb-4de4-476d-bf11-501d0c2a10b2", 00:13:43.432 "strip_size_kb": 0, 00:13:43.432 "state": "online", 00:13:43.432 "raid_level": "raid1", 00:13:43.432 "superblock": false, 00:13:43.432 "num_base_bdevs": 4, 00:13:43.432 "num_base_bdevs_discovered": 3, 00:13:43.432 "num_base_bdevs_operational": 3, 00:13:43.432 "base_bdevs_list": [ 00:13:43.432 { 00:13:43.432 "name": null, 00:13:43.432 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:43.432 "is_configured": false, 00:13:43.432 "data_offset": 0, 00:13:43.432 "data_size": 65536 00:13:43.432 }, 00:13:43.432 { 00:13:43.432 "name": "BaseBdev2", 00:13:43.432 "uuid": "154a7b63-1ad7-5526-8b8f-8904d1af1a44", 00:13:43.432 "is_configured": true, 00:13:43.432 "data_offset": 0, 00:13:43.432 "data_size": 65536 00:13:43.432 }, 00:13:43.432 { 00:13:43.432 "name": "BaseBdev3", 00:13:43.432 "uuid": "9a487f7d-0585-539a-98e3-1f90b2da8abd", 00:13:43.432 "is_configured": true, 00:13:43.432 "data_offset": 0, 00:13:43.432 "data_size": 65536 00:13:43.432 }, 00:13:43.432 { 00:13:43.432 "name": "BaseBdev4", 00:13:43.432 "uuid": "36cfb9f6-2877-5bae-90d5-98d75414362e", 00:13:43.432 "is_configured": true, 00:13:43.432 "data_offset": 0, 00:13:43.432 "data_size": 65536 00:13:43.432 } 00:13:43.432 ] 00:13:43.432 }' 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:43.432 16:01:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.742 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:43.742 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:43.742 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:43.742 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:43.742 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:43.742 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.742 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:43.742 16:01:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.742 16:01:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.742 16:01:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.742 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:43.742 "name": "raid_bdev1", 00:13:43.742 "uuid": "abb7ffcb-4de4-476d-bf11-501d0c2a10b2", 00:13:43.742 "strip_size_kb": 0, 00:13:43.742 "state": "online", 00:13:43.742 "raid_level": "raid1", 00:13:43.742 "superblock": false, 00:13:43.742 "num_base_bdevs": 4, 00:13:43.742 "num_base_bdevs_discovered": 3, 00:13:43.742 "num_base_bdevs_operational": 3, 00:13:43.742 "base_bdevs_list": [ 00:13:43.742 { 00:13:43.742 "name": null, 00:13:43.742 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:43.742 "is_configured": false, 00:13:43.742 "data_offset": 0, 00:13:43.742 "data_size": 65536 00:13:43.742 }, 00:13:43.742 { 00:13:43.742 "name": "BaseBdev2", 00:13:43.742 "uuid": "154a7b63-1ad7-5526-8b8f-8904d1af1a44", 00:13:43.742 "is_configured": true, 00:13:43.742 "data_offset": 0, 00:13:43.742 "data_size": 65536 00:13:43.742 }, 00:13:43.743 { 00:13:43.743 "name": "BaseBdev3", 00:13:43.743 "uuid": "9a487f7d-0585-539a-98e3-1f90b2da8abd", 00:13:43.743 "is_configured": true, 00:13:43.743 "data_offset": 0, 00:13:43.743 "data_size": 65536 00:13:43.743 }, 00:13:43.743 { 00:13:43.743 "name": "BaseBdev4", 00:13:43.743 "uuid": "36cfb9f6-2877-5bae-90d5-98d75414362e", 00:13:43.743 "is_configured": true, 00:13:43.743 "data_offset": 0, 00:13:43.743 "data_size": 65536 00:13:43.743 } 00:13:43.743 ] 00:13:43.743 }' 00:13:43.743 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:43.743 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:43.743 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:43.743 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:43.743 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:43.743 16:01:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.743 16:01:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.743 [2024-11-20 16:01:41.880159] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:43.743 [2024-11-20 16:01:41.887952] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09e40 00:13:43.743 16:01:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.743 16:01:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:43.743 [2024-11-20 16:01:41.889638] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:44.675 16:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:44.675 16:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:44.675 16:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:44.675 16:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:44.675 16:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:44.675 16:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:44.675 16:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:44.675 16:01:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.675 16:01:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:44.675 16:01:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.933 16:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:44.933 "name": "raid_bdev1", 00:13:44.933 "uuid": "abb7ffcb-4de4-476d-bf11-501d0c2a10b2", 00:13:44.933 "strip_size_kb": 0, 00:13:44.933 "state": "online", 00:13:44.933 "raid_level": "raid1", 00:13:44.933 "superblock": false, 00:13:44.933 "num_base_bdevs": 4, 00:13:44.933 "num_base_bdevs_discovered": 4, 00:13:44.933 "num_base_bdevs_operational": 4, 00:13:44.933 "process": { 00:13:44.933 "type": "rebuild", 00:13:44.933 "target": "spare", 00:13:44.933 "progress": { 00:13:44.933 "blocks": 20480, 00:13:44.933 "percent": 31 00:13:44.933 } 00:13:44.933 }, 00:13:44.933 "base_bdevs_list": [ 00:13:44.933 { 00:13:44.933 "name": "spare", 00:13:44.933 "uuid": "ea39c658-7bc1-5a69-b45e-bffc31a97f83", 00:13:44.933 "is_configured": true, 00:13:44.933 "data_offset": 0, 00:13:44.933 "data_size": 65536 00:13:44.933 }, 00:13:44.933 { 00:13:44.933 "name": "BaseBdev2", 00:13:44.933 "uuid": "154a7b63-1ad7-5526-8b8f-8904d1af1a44", 00:13:44.933 "is_configured": true, 00:13:44.933 "data_offset": 0, 00:13:44.933 "data_size": 65536 00:13:44.933 }, 00:13:44.933 { 00:13:44.933 "name": "BaseBdev3", 00:13:44.933 "uuid": "9a487f7d-0585-539a-98e3-1f90b2da8abd", 00:13:44.933 "is_configured": true, 00:13:44.933 "data_offset": 0, 00:13:44.933 "data_size": 65536 00:13:44.933 }, 00:13:44.933 { 00:13:44.933 "name": "BaseBdev4", 00:13:44.933 "uuid": "36cfb9f6-2877-5bae-90d5-98d75414362e", 00:13:44.933 "is_configured": true, 00:13:44.933 "data_offset": 0, 00:13:44.933 "data_size": 65536 00:13:44.933 } 00:13:44.933 ] 00:13:44.933 }' 00:13:44.933 16:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:44.933 16:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:44.933 16:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:44.933 16:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:44.933 16:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:13:44.933 16:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:13:44.933 16:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:13:44.933 16:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:13:44.933 16:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:44.933 16:01:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.933 16:01:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:44.933 [2024-11-20 16:01:42.995498] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:44.933 [2024-11-20 16:01:43.095080] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000d09e40 00:13:44.934 16:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.934 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:13:44.934 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:13:44.934 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:44.934 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:44.934 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:44.934 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:44.934 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:44.934 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:44.934 16:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.934 16:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:44.934 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:44.934 16:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.934 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:44.934 "name": "raid_bdev1", 00:13:44.934 "uuid": "abb7ffcb-4de4-476d-bf11-501d0c2a10b2", 00:13:44.934 "strip_size_kb": 0, 00:13:44.934 "state": "online", 00:13:44.934 "raid_level": "raid1", 00:13:44.934 "superblock": false, 00:13:44.934 "num_base_bdevs": 4, 00:13:44.934 "num_base_bdevs_discovered": 3, 00:13:44.934 "num_base_bdevs_operational": 3, 00:13:44.934 "process": { 00:13:44.934 "type": "rebuild", 00:13:44.934 "target": "spare", 00:13:44.934 "progress": { 00:13:44.934 "blocks": 24576, 00:13:44.934 "percent": 37 00:13:44.934 } 00:13:44.934 }, 00:13:44.934 "base_bdevs_list": [ 00:13:44.934 { 00:13:44.934 "name": "spare", 00:13:44.934 "uuid": "ea39c658-7bc1-5a69-b45e-bffc31a97f83", 00:13:44.934 "is_configured": true, 00:13:44.934 "data_offset": 0, 00:13:44.934 "data_size": 65536 00:13:44.934 }, 00:13:44.934 { 00:13:44.934 "name": null, 00:13:44.934 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:44.934 "is_configured": false, 00:13:44.934 "data_offset": 0, 00:13:44.934 "data_size": 65536 00:13:44.934 }, 00:13:44.934 { 00:13:44.934 "name": "BaseBdev3", 00:13:44.934 "uuid": "9a487f7d-0585-539a-98e3-1f90b2da8abd", 00:13:44.934 "is_configured": true, 00:13:44.934 "data_offset": 0, 00:13:44.934 "data_size": 65536 00:13:44.934 }, 00:13:44.934 { 00:13:44.934 "name": "BaseBdev4", 00:13:44.934 "uuid": "36cfb9f6-2877-5bae-90d5-98d75414362e", 00:13:44.934 "is_configured": true, 00:13:44.934 "data_offset": 0, 00:13:44.934 "data_size": 65536 00:13:44.934 } 00:13:44.934 ] 00:13:44.934 }' 00:13:44.934 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:44.934 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:44.934 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=357 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:45.191 "name": "raid_bdev1", 00:13:45.191 "uuid": "abb7ffcb-4de4-476d-bf11-501d0c2a10b2", 00:13:45.191 "strip_size_kb": 0, 00:13:45.191 "state": "online", 00:13:45.191 "raid_level": "raid1", 00:13:45.191 "superblock": false, 00:13:45.191 "num_base_bdevs": 4, 00:13:45.191 "num_base_bdevs_discovered": 3, 00:13:45.191 "num_base_bdevs_operational": 3, 00:13:45.191 "process": { 00:13:45.191 "type": "rebuild", 00:13:45.191 "target": "spare", 00:13:45.191 "progress": { 00:13:45.191 "blocks": 26624, 00:13:45.191 "percent": 40 00:13:45.191 } 00:13:45.191 }, 00:13:45.191 "base_bdevs_list": [ 00:13:45.191 { 00:13:45.191 "name": "spare", 00:13:45.191 "uuid": "ea39c658-7bc1-5a69-b45e-bffc31a97f83", 00:13:45.191 "is_configured": true, 00:13:45.191 "data_offset": 0, 00:13:45.191 "data_size": 65536 00:13:45.191 }, 00:13:45.191 { 00:13:45.191 "name": null, 00:13:45.191 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:45.191 "is_configured": false, 00:13:45.191 "data_offset": 0, 00:13:45.191 "data_size": 65536 00:13:45.191 }, 00:13:45.191 { 00:13:45.191 "name": "BaseBdev3", 00:13:45.191 "uuid": "9a487f7d-0585-539a-98e3-1f90b2da8abd", 00:13:45.191 "is_configured": true, 00:13:45.191 "data_offset": 0, 00:13:45.191 "data_size": 65536 00:13:45.191 }, 00:13:45.191 { 00:13:45.191 "name": "BaseBdev4", 00:13:45.191 "uuid": "36cfb9f6-2877-5bae-90d5-98d75414362e", 00:13:45.191 "is_configured": true, 00:13:45.191 "data_offset": 0, 00:13:45.191 "data_size": 65536 00:13:45.191 } 00:13:45.191 ] 00:13:45.191 }' 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:45.191 16:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:46.123 16:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:46.123 16:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:46.123 16:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:46.123 16:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:46.123 16:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:46.123 16:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:46.123 16:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.124 16:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:46.124 16:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.124 16:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:46.124 16:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.124 16:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:46.124 "name": "raid_bdev1", 00:13:46.124 "uuid": "abb7ffcb-4de4-476d-bf11-501d0c2a10b2", 00:13:46.124 "strip_size_kb": 0, 00:13:46.124 "state": "online", 00:13:46.124 "raid_level": "raid1", 00:13:46.124 "superblock": false, 00:13:46.124 "num_base_bdevs": 4, 00:13:46.124 "num_base_bdevs_discovered": 3, 00:13:46.124 "num_base_bdevs_operational": 3, 00:13:46.124 "process": { 00:13:46.124 "type": "rebuild", 00:13:46.124 "target": "spare", 00:13:46.124 "progress": { 00:13:46.124 "blocks": 47104, 00:13:46.124 "percent": 71 00:13:46.124 } 00:13:46.124 }, 00:13:46.124 "base_bdevs_list": [ 00:13:46.124 { 00:13:46.124 "name": "spare", 00:13:46.124 "uuid": "ea39c658-7bc1-5a69-b45e-bffc31a97f83", 00:13:46.124 "is_configured": true, 00:13:46.124 "data_offset": 0, 00:13:46.124 "data_size": 65536 00:13:46.124 }, 00:13:46.124 { 00:13:46.124 "name": null, 00:13:46.124 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:46.124 "is_configured": false, 00:13:46.124 "data_offset": 0, 00:13:46.124 "data_size": 65536 00:13:46.124 }, 00:13:46.124 { 00:13:46.124 "name": "BaseBdev3", 00:13:46.124 "uuid": "9a487f7d-0585-539a-98e3-1f90b2da8abd", 00:13:46.124 "is_configured": true, 00:13:46.124 "data_offset": 0, 00:13:46.124 "data_size": 65536 00:13:46.124 }, 00:13:46.124 { 00:13:46.124 "name": "BaseBdev4", 00:13:46.124 "uuid": "36cfb9f6-2877-5bae-90d5-98d75414362e", 00:13:46.124 "is_configured": true, 00:13:46.124 "data_offset": 0, 00:13:46.124 "data_size": 65536 00:13:46.124 } 00:13:46.124 ] 00:13:46.124 }' 00:13:46.124 16:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:46.124 16:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:46.124 16:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:46.382 16:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:46.382 16:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:46.947 [2024-11-20 16:01:45.104137] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:46.947 [2024-11-20 16:01:45.104211] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:46.947 [2024-11-20 16:01:45.104253] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:47.204 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:47.204 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:47.204 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:47.204 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:47.204 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:47.204 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:47.204 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:47.204 16:01:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.204 16:01:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.204 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:47.204 16:01:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.204 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:47.204 "name": "raid_bdev1", 00:13:47.204 "uuid": "abb7ffcb-4de4-476d-bf11-501d0c2a10b2", 00:13:47.204 "strip_size_kb": 0, 00:13:47.204 "state": "online", 00:13:47.204 "raid_level": "raid1", 00:13:47.204 "superblock": false, 00:13:47.204 "num_base_bdevs": 4, 00:13:47.204 "num_base_bdevs_discovered": 3, 00:13:47.204 "num_base_bdevs_operational": 3, 00:13:47.204 "base_bdevs_list": [ 00:13:47.204 { 00:13:47.204 "name": "spare", 00:13:47.204 "uuid": "ea39c658-7bc1-5a69-b45e-bffc31a97f83", 00:13:47.204 "is_configured": true, 00:13:47.204 "data_offset": 0, 00:13:47.204 "data_size": 65536 00:13:47.204 }, 00:13:47.204 { 00:13:47.204 "name": null, 00:13:47.204 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:47.204 "is_configured": false, 00:13:47.204 "data_offset": 0, 00:13:47.204 "data_size": 65536 00:13:47.204 }, 00:13:47.204 { 00:13:47.204 "name": "BaseBdev3", 00:13:47.204 "uuid": "9a487f7d-0585-539a-98e3-1f90b2da8abd", 00:13:47.204 "is_configured": true, 00:13:47.204 "data_offset": 0, 00:13:47.204 "data_size": 65536 00:13:47.204 }, 00:13:47.204 { 00:13:47.204 "name": "BaseBdev4", 00:13:47.204 "uuid": "36cfb9f6-2877-5bae-90d5-98d75414362e", 00:13:47.205 "is_configured": true, 00:13:47.205 "data_offset": 0, 00:13:47.205 "data_size": 65536 00:13:47.205 } 00:13:47.205 ] 00:13:47.205 }' 00:13:47.205 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:47.205 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:47.205 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:47.463 "name": "raid_bdev1", 00:13:47.463 "uuid": "abb7ffcb-4de4-476d-bf11-501d0c2a10b2", 00:13:47.463 "strip_size_kb": 0, 00:13:47.463 "state": "online", 00:13:47.463 "raid_level": "raid1", 00:13:47.463 "superblock": false, 00:13:47.463 "num_base_bdevs": 4, 00:13:47.463 "num_base_bdevs_discovered": 3, 00:13:47.463 "num_base_bdevs_operational": 3, 00:13:47.463 "base_bdevs_list": [ 00:13:47.463 { 00:13:47.463 "name": "spare", 00:13:47.463 "uuid": "ea39c658-7bc1-5a69-b45e-bffc31a97f83", 00:13:47.463 "is_configured": true, 00:13:47.463 "data_offset": 0, 00:13:47.463 "data_size": 65536 00:13:47.463 }, 00:13:47.463 { 00:13:47.463 "name": null, 00:13:47.463 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:47.463 "is_configured": false, 00:13:47.463 "data_offset": 0, 00:13:47.463 "data_size": 65536 00:13:47.463 }, 00:13:47.463 { 00:13:47.463 "name": "BaseBdev3", 00:13:47.463 "uuid": "9a487f7d-0585-539a-98e3-1f90b2da8abd", 00:13:47.463 "is_configured": true, 00:13:47.463 "data_offset": 0, 00:13:47.463 "data_size": 65536 00:13:47.463 }, 00:13:47.463 { 00:13:47.463 "name": "BaseBdev4", 00:13:47.463 "uuid": "36cfb9f6-2877-5bae-90d5-98d75414362e", 00:13:47.463 "is_configured": true, 00:13:47.463 "data_offset": 0, 00:13:47.463 "data_size": 65536 00:13:47.463 } 00:13:47.463 ] 00:13:47.463 }' 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:47.463 "name": "raid_bdev1", 00:13:47.463 "uuid": "abb7ffcb-4de4-476d-bf11-501d0c2a10b2", 00:13:47.463 "strip_size_kb": 0, 00:13:47.463 "state": "online", 00:13:47.463 "raid_level": "raid1", 00:13:47.463 "superblock": false, 00:13:47.463 "num_base_bdevs": 4, 00:13:47.463 "num_base_bdevs_discovered": 3, 00:13:47.463 "num_base_bdevs_operational": 3, 00:13:47.463 "base_bdevs_list": [ 00:13:47.463 { 00:13:47.463 "name": "spare", 00:13:47.463 "uuid": "ea39c658-7bc1-5a69-b45e-bffc31a97f83", 00:13:47.463 "is_configured": true, 00:13:47.463 "data_offset": 0, 00:13:47.463 "data_size": 65536 00:13:47.463 }, 00:13:47.463 { 00:13:47.463 "name": null, 00:13:47.463 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:47.463 "is_configured": false, 00:13:47.463 "data_offset": 0, 00:13:47.463 "data_size": 65536 00:13:47.463 }, 00:13:47.463 { 00:13:47.463 "name": "BaseBdev3", 00:13:47.463 "uuid": "9a487f7d-0585-539a-98e3-1f90b2da8abd", 00:13:47.463 "is_configured": true, 00:13:47.463 "data_offset": 0, 00:13:47.463 "data_size": 65536 00:13:47.463 }, 00:13:47.463 { 00:13:47.463 "name": "BaseBdev4", 00:13:47.463 "uuid": "36cfb9f6-2877-5bae-90d5-98d75414362e", 00:13:47.463 "is_configured": true, 00:13:47.463 "data_offset": 0, 00:13:47.463 "data_size": 65536 00:13:47.463 } 00:13:47.463 ] 00:13:47.463 }' 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:47.463 16:01:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.721 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:47.721 16:01:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.721 16:01:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.721 [2024-11-20 16:01:45.888412] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:47.722 [2024-11-20 16:01:45.888528] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:47.722 [2024-11-20 16:01:45.888598] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:47.722 [2024-11-20 16:01:45.888665] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:47.722 [2024-11-20 16:01:45.888687] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:47.722 16:01:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:13:47.980 /dev/nbd0 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:47.980 1+0 records in 00:13:47.980 1+0 records out 00:13:47.980 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00028552 s, 14.3 MB/s 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:47.980 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:13:48.239 /dev/nbd1 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:48.239 1+0 records in 00:13:48.239 1+0 records out 00:13:48.239 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000363517 s, 11.3 MB/s 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:48.239 16:01:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:13:48.496 16:01:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:13:48.496 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:48.496 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:48.496 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:48.496 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:13:48.496 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:48.496 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:48.496 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:48.496 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:48.496 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:48.496 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:48.496 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:48.496 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:48.496 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:48.496 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:48.496 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:48.496 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 75417 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 75417 ']' 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 75417 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 75417 00:13:48.755 killing process with pid 75417 00:13:48.755 Received shutdown signal, test time was about 60.000000 seconds 00:13:48.755 00:13:48.755 Latency(us) 00:13:48.755 [2024-11-20T16:01:47.006Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:48.755 [2024-11-20T16:01:47.006Z] =================================================================================================================== 00:13:48.755 [2024-11-20T16:01:47.006Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 75417' 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@973 -- # kill 75417 00:13:48.755 [2024-11-20 16:01:46.967487] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:48.755 16:01:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@978 -- # wait 75417 00:13:49.012 [2024-11-20 16:01:47.207963] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:49.580 ************************************ 00:13:49.580 END TEST raid_rebuild_test 00:13:49.580 ************************************ 00:13:49.580 16:01:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:13:49.580 00:13:49.580 real 0m18.319s 00:13:49.580 user 0m19.404s 00:13:49.580 sys 0m3.424s 00:13:49.580 16:01:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:49.580 16:01:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:49.841 16:01:47 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 4 true false true 00:13:49.841 16:01:47 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:13:49.841 16:01:47 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:49.841 16:01:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:49.841 ************************************ 00:13:49.841 START TEST raid_rebuild_test_sb 00:13:49.841 ************************************ 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 true false true 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:13:49.841 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=75874 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 75874 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 75874 ']' 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:49.841 16:01:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.841 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:49.841 Zero copy mechanism will not be used. 00:13:49.841 [2024-11-20 16:01:47.912327] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:13:49.841 [2024-11-20 16:01:47.912451] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid75874 ] 00:13:49.841 [2024-11-20 16:01:48.070301] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:50.102 [2024-11-20 16:01:48.172953] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:50.102 [2024-11-20 16:01:48.309746] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:50.102 [2024-11-20 16:01:48.309944] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.672 BaseBdev1_malloc 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.672 [2024-11-20 16:01:48.792489] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:50.672 [2024-11-20 16:01:48.792656] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:50.672 [2024-11-20 16:01:48.792712] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:50.672 [2024-11-20 16:01:48.792773] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:50.672 [2024-11-20 16:01:48.794948] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:50.672 [2024-11-20 16:01:48.795068] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:50.672 BaseBdev1 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.672 BaseBdev2_malloc 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.672 [2024-11-20 16:01:48.832248] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:50.672 [2024-11-20 16:01:48.832300] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:50.672 [2024-11-20 16:01:48.832320] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:50.672 [2024-11-20 16:01:48.832329] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:50.672 [2024-11-20 16:01:48.834416] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:50.672 [2024-11-20 16:01:48.834451] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:50.672 BaseBdev2 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.672 BaseBdev3_malloc 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.672 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.672 [2024-11-20 16:01:48.875203] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:50.672 [2024-11-20 16:01:48.875355] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:50.672 [2024-11-20 16:01:48.875383] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:50.672 [2024-11-20 16:01:48.875394] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:50.673 [2024-11-20 16:01:48.877502] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:50.673 [2024-11-20 16:01:48.877539] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:50.673 BaseBdev3 00:13:50.673 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.673 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:50.673 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:13:50.673 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.673 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.673 BaseBdev4_malloc 00:13:50.673 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.673 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:13:50.673 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.673 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.673 [2024-11-20 16:01:48.915284] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:13:50.673 [2024-11-20 16:01:48.915422] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:50.673 [2024-11-20 16:01:48.915458] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:13:50.673 [2024-11-20 16:01:48.915510] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:50.673 [2024-11-20 16:01:48.917559] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:50.673 [2024-11-20 16:01:48.917675] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:13:50.673 BaseBdev4 00:13:50.673 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.673 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:50.673 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.673 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.934 spare_malloc 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.934 spare_delay 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.934 [2024-11-20 16:01:48.963038] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:50.934 [2024-11-20 16:01:48.963175] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:50.934 [2024-11-20 16:01:48.963213] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:13:50.934 [2024-11-20 16:01:48.963264] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:50.934 [2024-11-20 16:01:48.965371] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:50.934 [2024-11-20 16:01:48.965479] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:50.934 spare 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.934 [2024-11-20 16:01:48.971092] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:50.934 [2024-11-20 16:01:48.972980] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:50.934 [2024-11-20 16:01:48.973042] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:50.934 [2024-11-20 16:01:48.973094] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:50.934 [2024-11-20 16:01:48.973273] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:13:50.934 [2024-11-20 16:01:48.973286] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:50.934 [2024-11-20 16:01:48.973536] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:13:50.934 [2024-11-20 16:01:48.973701] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:13:50.934 [2024-11-20 16:01:48.973711] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:13:50.934 [2024-11-20 16:01:48.973850] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.934 16:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.934 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:50.934 "name": "raid_bdev1", 00:13:50.934 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:13:50.934 "strip_size_kb": 0, 00:13:50.934 "state": "online", 00:13:50.934 "raid_level": "raid1", 00:13:50.934 "superblock": true, 00:13:50.934 "num_base_bdevs": 4, 00:13:50.934 "num_base_bdevs_discovered": 4, 00:13:50.934 "num_base_bdevs_operational": 4, 00:13:50.934 "base_bdevs_list": [ 00:13:50.934 { 00:13:50.934 "name": "BaseBdev1", 00:13:50.934 "uuid": "15d4d1e4-401e-57e9-908f-9825a7b73eba", 00:13:50.934 "is_configured": true, 00:13:50.934 "data_offset": 2048, 00:13:50.934 "data_size": 63488 00:13:50.934 }, 00:13:50.934 { 00:13:50.934 "name": "BaseBdev2", 00:13:50.934 "uuid": "c4eee168-194e-54f9-927b-48bf1da0146a", 00:13:50.934 "is_configured": true, 00:13:50.934 "data_offset": 2048, 00:13:50.934 "data_size": 63488 00:13:50.934 }, 00:13:50.934 { 00:13:50.934 "name": "BaseBdev3", 00:13:50.934 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:13:50.934 "is_configured": true, 00:13:50.934 "data_offset": 2048, 00:13:50.934 "data_size": 63488 00:13:50.934 }, 00:13:50.934 { 00:13:50.934 "name": "BaseBdev4", 00:13:50.934 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:13:50.934 "is_configured": true, 00:13:50.934 "data_offset": 2048, 00:13:50.934 "data_size": 63488 00:13:50.934 } 00:13:50.934 ] 00:13:50.934 }' 00:13:50.934 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:50.934 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:51.195 [2024-11-20 16:01:49.307537] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:51.195 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:51.196 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:13:51.196 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:51.196 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:51.196 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:13:51.457 [2024-11-20 16:01:49.547268] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:13:51.457 /dev/nbd0 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:51.457 1+0 records in 00:13:51.457 1+0 records out 00:13:51.457 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000426908 s, 9.6 MB/s 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:13:51.457 16:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:13:59.642 63488+0 records in 00:13:59.642 63488+0 records out 00:13:59.642 32505856 bytes (33 MB, 31 MiB) copied, 7.19448 s, 4.5 MB/s 00:13:59.642 16:01:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:59.642 16:01:56 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:59.642 16:01:56 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:59.642 16:01:56 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:59.642 16:01:56 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:13:59.643 16:01:56 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:59.643 16:01:56 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:59.643 [2024-11-20 16:01:56.990869] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.643 [2024-11-20 16:01:57.018938] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:59.643 "name": "raid_bdev1", 00:13:59.643 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:13:59.643 "strip_size_kb": 0, 00:13:59.643 "state": "online", 00:13:59.643 "raid_level": "raid1", 00:13:59.643 "superblock": true, 00:13:59.643 "num_base_bdevs": 4, 00:13:59.643 "num_base_bdevs_discovered": 3, 00:13:59.643 "num_base_bdevs_operational": 3, 00:13:59.643 "base_bdevs_list": [ 00:13:59.643 { 00:13:59.643 "name": null, 00:13:59.643 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:59.643 "is_configured": false, 00:13:59.643 "data_offset": 0, 00:13:59.643 "data_size": 63488 00:13:59.643 }, 00:13:59.643 { 00:13:59.643 "name": "BaseBdev2", 00:13:59.643 "uuid": "c4eee168-194e-54f9-927b-48bf1da0146a", 00:13:59.643 "is_configured": true, 00:13:59.643 "data_offset": 2048, 00:13:59.643 "data_size": 63488 00:13:59.643 }, 00:13:59.643 { 00:13:59.643 "name": "BaseBdev3", 00:13:59.643 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:13:59.643 "is_configured": true, 00:13:59.643 "data_offset": 2048, 00:13:59.643 "data_size": 63488 00:13:59.643 }, 00:13:59.643 { 00:13:59.643 "name": "BaseBdev4", 00:13:59.643 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:13:59.643 "is_configured": true, 00:13:59.643 "data_offset": 2048, 00:13:59.643 "data_size": 63488 00:13:59.643 } 00:13:59.643 ] 00:13:59.643 }' 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.643 [2024-11-20 16:01:57.339008] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:59.643 [2024-11-20 16:01:57.349099] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca3500 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.643 16:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:59.643 [2024-11-20 16:01:57.351005] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:00.207 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:00.207 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:00.207 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:00.207 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:00.207 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:00.208 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.208 16:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:00.208 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:00.208 16:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.208 16:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:00.208 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:00.208 "name": "raid_bdev1", 00:14:00.208 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:00.208 "strip_size_kb": 0, 00:14:00.208 "state": "online", 00:14:00.208 "raid_level": "raid1", 00:14:00.208 "superblock": true, 00:14:00.208 "num_base_bdevs": 4, 00:14:00.208 "num_base_bdevs_discovered": 4, 00:14:00.208 "num_base_bdevs_operational": 4, 00:14:00.208 "process": { 00:14:00.208 "type": "rebuild", 00:14:00.208 "target": "spare", 00:14:00.208 "progress": { 00:14:00.208 "blocks": 20480, 00:14:00.208 "percent": 32 00:14:00.208 } 00:14:00.208 }, 00:14:00.208 "base_bdevs_list": [ 00:14:00.208 { 00:14:00.208 "name": "spare", 00:14:00.208 "uuid": "763fe959-7afc-540e-ae2a-f141a41763f1", 00:14:00.208 "is_configured": true, 00:14:00.208 "data_offset": 2048, 00:14:00.208 "data_size": 63488 00:14:00.208 }, 00:14:00.208 { 00:14:00.208 "name": "BaseBdev2", 00:14:00.208 "uuid": "c4eee168-194e-54f9-927b-48bf1da0146a", 00:14:00.208 "is_configured": true, 00:14:00.208 "data_offset": 2048, 00:14:00.208 "data_size": 63488 00:14:00.208 }, 00:14:00.208 { 00:14:00.208 "name": "BaseBdev3", 00:14:00.208 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:00.208 "is_configured": true, 00:14:00.208 "data_offset": 2048, 00:14:00.208 "data_size": 63488 00:14:00.208 }, 00:14:00.208 { 00:14:00.208 "name": "BaseBdev4", 00:14:00.208 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:00.208 "is_configured": true, 00:14:00.208 "data_offset": 2048, 00:14:00.208 "data_size": 63488 00:14:00.208 } 00:14:00.208 ] 00:14:00.208 }' 00:14:00.208 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:00.208 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:00.208 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:00.208 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:00.208 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:00.208 16:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:00.208 16:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.465 [2024-11-20 16:01:58.460520] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:00.465 [2024-11-20 16:01:58.556416] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:00.465 [2024-11-20 16:01:58.556480] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:00.465 [2024-11-20 16:01:58.556494] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:00.465 [2024-11-20 16:01:58.556502] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:00.465 "name": "raid_bdev1", 00:14:00.465 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:00.465 "strip_size_kb": 0, 00:14:00.465 "state": "online", 00:14:00.465 "raid_level": "raid1", 00:14:00.465 "superblock": true, 00:14:00.465 "num_base_bdevs": 4, 00:14:00.465 "num_base_bdevs_discovered": 3, 00:14:00.465 "num_base_bdevs_operational": 3, 00:14:00.465 "base_bdevs_list": [ 00:14:00.465 { 00:14:00.465 "name": null, 00:14:00.465 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:00.465 "is_configured": false, 00:14:00.465 "data_offset": 0, 00:14:00.465 "data_size": 63488 00:14:00.465 }, 00:14:00.465 { 00:14:00.465 "name": "BaseBdev2", 00:14:00.465 "uuid": "c4eee168-194e-54f9-927b-48bf1da0146a", 00:14:00.465 "is_configured": true, 00:14:00.465 "data_offset": 2048, 00:14:00.465 "data_size": 63488 00:14:00.465 }, 00:14:00.465 { 00:14:00.465 "name": "BaseBdev3", 00:14:00.465 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:00.465 "is_configured": true, 00:14:00.465 "data_offset": 2048, 00:14:00.465 "data_size": 63488 00:14:00.465 }, 00:14:00.465 { 00:14:00.465 "name": "BaseBdev4", 00:14:00.465 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:00.465 "is_configured": true, 00:14:00.465 "data_offset": 2048, 00:14:00.465 "data_size": 63488 00:14:00.465 } 00:14:00.465 ] 00:14:00.465 }' 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:00.465 16:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.722 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:00.722 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:00.722 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:00.722 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:00.722 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:00.722 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.722 16:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:00.722 16:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.722 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:00.722 16:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:00.722 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:00.722 "name": "raid_bdev1", 00:14:00.722 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:00.722 "strip_size_kb": 0, 00:14:00.722 "state": "online", 00:14:00.722 "raid_level": "raid1", 00:14:00.722 "superblock": true, 00:14:00.722 "num_base_bdevs": 4, 00:14:00.722 "num_base_bdevs_discovered": 3, 00:14:00.722 "num_base_bdevs_operational": 3, 00:14:00.722 "base_bdevs_list": [ 00:14:00.722 { 00:14:00.722 "name": null, 00:14:00.722 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:00.722 "is_configured": false, 00:14:00.722 "data_offset": 0, 00:14:00.722 "data_size": 63488 00:14:00.722 }, 00:14:00.722 { 00:14:00.722 "name": "BaseBdev2", 00:14:00.722 "uuid": "c4eee168-194e-54f9-927b-48bf1da0146a", 00:14:00.722 "is_configured": true, 00:14:00.722 "data_offset": 2048, 00:14:00.722 "data_size": 63488 00:14:00.722 }, 00:14:00.722 { 00:14:00.722 "name": "BaseBdev3", 00:14:00.722 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:00.722 "is_configured": true, 00:14:00.722 "data_offset": 2048, 00:14:00.722 "data_size": 63488 00:14:00.722 }, 00:14:00.722 { 00:14:00.722 "name": "BaseBdev4", 00:14:00.722 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:00.722 "is_configured": true, 00:14:00.722 "data_offset": 2048, 00:14:00.722 "data_size": 63488 00:14:00.722 } 00:14:00.722 ] 00:14:00.722 }' 00:14:00.722 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:00.722 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:00.722 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:00.980 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:00.980 16:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:00.980 16:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:00.980 16:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.980 [2024-11-20 16:01:58.992603] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:00.980 [2024-11-20 16:01:59.000183] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca35d0 00:14:00.980 16:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:00.980 16:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:00.980 [2024-11-20 16:01:59.001809] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:01.912 "name": "raid_bdev1", 00:14:01.912 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:01.912 "strip_size_kb": 0, 00:14:01.912 "state": "online", 00:14:01.912 "raid_level": "raid1", 00:14:01.912 "superblock": true, 00:14:01.912 "num_base_bdevs": 4, 00:14:01.912 "num_base_bdevs_discovered": 4, 00:14:01.912 "num_base_bdevs_operational": 4, 00:14:01.912 "process": { 00:14:01.912 "type": "rebuild", 00:14:01.912 "target": "spare", 00:14:01.912 "progress": { 00:14:01.912 "blocks": 20480, 00:14:01.912 "percent": 32 00:14:01.912 } 00:14:01.912 }, 00:14:01.912 "base_bdevs_list": [ 00:14:01.912 { 00:14:01.912 "name": "spare", 00:14:01.912 "uuid": "763fe959-7afc-540e-ae2a-f141a41763f1", 00:14:01.912 "is_configured": true, 00:14:01.912 "data_offset": 2048, 00:14:01.912 "data_size": 63488 00:14:01.912 }, 00:14:01.912 { 00:14:01.912 "name": "BaseBdev2", 00:14:01.912 "uuid": "c4eee168-194e-54f9-927b-48bf1da0146a", 00:14:01.912 "is_configured": true, 00:14:01.912 "data_offset": 2048, 00:14:01.912 "data_size": 63488 00:14:01.912 }, 00:14:01.912 { 00:14:01.912 "name": "BaseBdev3", 00:14:01.912 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:01.912 "is_configured": true, 00:14:01.912 "data_offset": 2048, 00:14:01.912 "data_size": 63488 00:14:01.912 }, 00:14:01.912 { 00:14:01.912 "name": "BaseBdev4", 00:14:01.912 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:01.912 "is_configured": true, 00:14:01.912 "data_offset": 2048, 00:14:01.912 "data_size": 63488 00:14:01.912 } 00:14:01.912 ] 00:14:01.912 }' 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:14:01.912 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.912 16:02:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.912 [2024-11-20 16:02:00.107857] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:02.171 [2024-11-20 16:02:00.307005] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000ca35d0 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:02.171 "name": "raid_bdev1", 00:14:02.171 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:02.171 "strip_size_kb": 0, 00:14:02.171 "state": "online", 00:14:02.171 "raid_level": "raid1", 00:14:02.171 "superblock": true, 00:14:02.171 "num_base_bdevs": 4, 00:14:02.171 "num_base_bdevs_discovered": 3, 00:14:02.171 "num_base_bdevs_operational": 3, 00:14:02.171 "process": { 00:14:02.171 "type": "rebuild", 00:14:02.171 "target": "spare", 00:14:02.171 "progress": { 00:14:02.171 "blocks": 24576, 00:14:02.171 "percent": 38 00:14:02.171 } 00:14:02.171 }, 00:14:02.171 "base_bdevs_list": [ 00:14:02.171 { 00:14:02.171 "name": "spare", 00:14:02.171 "uuid": "763fe959-7afc-540e-ae2a-f141a41763f1", 00:14:02.171 "is_configured": true, 00:14:02.171 "data_offset": 2048, 00:14:02.171 "data_size": 63488 00:14:02.171 }, 00:14:02.171 { 00:14:02.171 "name": null, 00:14:02.171 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:02.171 "is_configured": false, 00:14:02.171 "data_offset": 0, 00:14:02.171 "data_size": 63488 00:14:02.171 }, 00:14:02.171 { 00:14:02.171 "name": "BaseBdev3", 00:14:02.171 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:02.171 "is_configured": true, 00:14:02.171 "data_offset": 2048, 00:14:02.171 "data_size": 63488 00:14:02.171 }, 00:14:02.171 { 00:14:02.171 "name": "BaseBdev4", 00:14:02.171 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:02.171 "is_configured": true, 00:14:02.171 "data_offset": 2048, 00:14:02.171 "data_size": 63488 00:14:02.171 } 00:14:02.171 ] 00:14:02.171 }' 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=374 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.171 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:02.428 16:02:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.428 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:02.428 "name": "raid_bdev1", 00:14:02.428 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:02.428 "strip_size_kb": 0, 00:14:02.429 "state": "online", 00:14:02.429 "raid_level": "raid1", 00:14:02.429 "superblock": true, 00:14:02.429 "num_base_bdevs": 4, 00:14:02.429 "num_base_bdevs_discovered": 3, 00:14:02.429 "num_base_bdevs_operational": 3, 00:14:02.429 "process": { 00:14:02.429 "type": "rebuild", 00:14:02.429 "target": "spare", 00:14:02.429 "progress": { 00:14:02.429 "blocks": 26624, 00:14:02.429 "percent": 41 00:14:02.429 } 00:14:02.429 }, 00:14:02.429 "base_bdevs_list": [ 00:14:02.429 { 00:14:02.429 "name": "spare", 00:14:02.429 "uuid": "763fe959-7afc-540e-ae2a-f141a41763f1", 00:14:02.429 "is_configured": true, 00:14:02.429 "data_offset": 2048, 00:14:02.429 "data_size": 63488 00:14:02.429 }, 00:14:02.429 { 00:14:02.429 "name": null, 00:14:02.429 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:02.429 "is_configured": false, 00:14:02.429 "data_offset": 0, 00:14:02.429 "data_size": 63488 00:14:02.429 }, 00:14:02.429 { 00:14:02.429 "name": "BaseBdev3", 00:14:02.429 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:02.429 "is_configured": true, 00:14:02.429 "data_offset": 2048, 00:14:02.429 "data_size": 63488 00:14:02.429 }, 00:14:02.429 { 00:14:02.429 "name": "BaseBdev4", 00:14:02.429 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:02.429 "is_configured": true, 00:14:02.429 "data_offset": 2048, 00:14:02.429 "data_size": 63488 00:14:02.429 } 00:14:02.429 ] 00:14:02.429 }' 00:14:02.429 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:02.429 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:02.429 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:02.429 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:02.429 16:02:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:03.359 16:02:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:03.359 16:02:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:03.359 16:02:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:03.359 16:02:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:03.359 16:02:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:03.359 16:02:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:03.359 16:02:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:03.359 16:02:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:03.359 16:02:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:03.359 16:02:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.359 16:02:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:03.359 16:02:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:03.359 "name": "raid_bdev1", 00:14:03.359 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:03.359 "strip_size_kb": 0, 00:14:03.360 "state": "online", 00:14:03.360 "raid_level": "raid1", 00:14:03.360 "superblock": true, 00:14:03.360 "num_base_bdevs": 4, 00:14:03.360 "num_base_bdevs_discovered": 3, 00:14:03.360 "num_base_bdevs_operational": 3, 00:14:03.360 "process": { 00:14:03.360 "type": "rebuild", 00:14:03.360 "target": "spare", 00:14:03.360 "progress": { 00:14:03.360 "blocks": 47104, 00:14:03.360 "percent": 74 00:14:03.360 } 00:14:03.360 }, 00:14:03.360 "base_bdevs_list": [ 00:14:03.360 { 00:14:03.360 "name": "spare", 00:14:03.360 "uuid": "763fe959-7afc-540e-ae2a-f141a41763f1", 00:14:03.360 "is_configured": true, 00:14:03.360 "data_offset": 2048, 00:14:03.360 "data_size": 63488 00:14:03.360 }, 00:14:03.360 { 00:14:03.360 "name": null, 00:14:03.360 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:03.360 "is_configured": false, 00:14:03.360 "data_offset": 0, 00:14:03.360 "data_size": 63488 00:14:03.360 }, 00:14:03.360 { 00:14:03.360 "name": "BaseBdev3", 00:14:03.360 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:03.360 "is_configured": true, 00:14:03.360 "data_offset": 2048, 00:14:03.360 "data_size": 63488 00:14:03.360 }, 00:14:03.360 { 00:14:03.360 "name": "BaseBdev4", 00:14:03.360 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:03.360 "is_configured": true, 00:14:03.360 "data_offset": 2048, 00:14:03.360 "data_size": 63488 00:14:03.360 } 00:14:03.360 ] 00:14:03.360 }' 00:14:03.360 16:02:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:03.360 16:02:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:03.360 16:02:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:03.360 16:02:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:03.360 16:02:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:04.292 [2024-11-20 16:02:02.215193] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:04.292 [2024-11-20 16:02:02.215257] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:04.292 [2024-11-20 16:02:02.215357] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:04.550 "name": "raid_bdev1", 00:14:04.550 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:04.550 "strip_size_kb": 0, 00:14:04.550 "state": "online", 00:14:04.550 "raid_level": "raid1", 00:14:04.550 "superblock": true, 00:14:04.550 "num_base_bdevs": 4, 00:14:04.550 "num_base_bdevs_discovered": 3, 00:14:04.550 "num_base_bdevs_operational": 3, 00:14:04.550 "base_bdevs_list": [ 00:14:04.550 { 00:14:04.550 "name": "spare", 00:14:04.550 "uuid": "763fe959-7afc-540e-ae2a-f141a41763f1", 00:14:04.550 "is_configured": true, 00:14:04.550 "data_offset": 2048, 00:14:04.550 "data_size": 63488 00:14:04.550 }, 00:14:04.550 { 00:14:04.550 "name": null, 00:14:04.550 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:04.550 "is_configured": false, 00:14:04.550 "data_offset": 0, 00:14:04.550 "data_size": 63488 00:14:04.550 }, 00:14:04.550 { 00:14:04.550 "name": "BaseBdev3", 00:14:04.550 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:04.550 "is_configured": true, 00:14:04.550 "data_offset": 2048, 00:14:04.550 "data_size": 63488 00:14:04.550 }, 00:14:04.550 { 00:14:04.550 "name": "BaseBdev4", 00:14:04.550 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:04.550 "is_configured": true, 00:14:04.550 "data_offset": 2048, 00:14:04.550 "data_size": 63488 00:14:04.550 } 00:14:04.550 ] 00:14:04.550 }' 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:04.550 "name": "raid_bdev1", 00:14:04.550 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:04.550 "strip_size_kb": 0, 00:14:04.550 "state": "online", 00:14:04.550 "raid_level": "raid1", 00:14:04.550 "superblock": true, 00:14:04.550 "num_base_bdevs": 4, 00:14:04.550 "num_base_bdevs_discovered": 3, 00:14:04.550 "num_base_bdevs_operational": 3, 00:14:04.550 "base_bdevs_list": [ 00:14:04.550 { 00:14:04.550 "name": "spare", 00:14:04.550 "uuid": "763fe959-7afc-540e-ae2a-f141a41763f1", 00:14:04.550 "is_configured": true, 00:14:04.550 "data_offset": 2048, 00:14:04.550 "data_size": 63488 00:14:04.550 }, 00:14:04.550 { 00:14:04.550 "name": null, 00:14:04.550 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:04.550 "is_configured": false, 00:14:04.550 "data_offset": 0, 00:14:04.550 "data_size": 63488 00:14:04.550 }, 00:14:04.550 { 00:14:04.550 "name": "BaseBdev3", 00:14:04.550 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:04.550 "is_configured": true, 00:14:04.550 "data_offset": 2048, 00:14:04.550 "data_size": 63488 00:14:04.550 }, 00:14:04.550 { 00:14:04.550 "name": "BaseBdev4", 00:14:04.550 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:04.550 "is_configured": true, 00:14:04.550 "data_offset": 2048, 00:14:04.550 "data_size": 63488 00:14:04.550 } 00:14:04.550 ] 00:14:04.550 }' 00:14:04.550 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:04.551 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:04.551 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:04.551 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:04.551 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:14:04.551 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:04.551 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:04.551 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:04.551 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:04.551 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:04.551 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:04.551 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:04.551 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:04.551 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:04.551 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.551 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:04.551 16:02:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.551 16:02:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.808 16:02:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.808 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:04.808 "name": "raid_bdev1", 00:14:04.808 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:04.808 "strip_size_kb": 0, 00:14:04.808 "state": "online", 00:14:04.808 "raid_level": "raid1", 00:14:04.808 "superblock": true, 00:14:04.808 "num_base_bdevs": 4, 00:14:04.808 "num_base_bdevs_discovered": 3, 00:14:04.808 "num_base_bdevs_operational": 3, 00:14:04.808 "base_bdevs_list": [ 00:14:04.808 { 00:14:04.808 "name": "spare", 00:14:04.808 "uuid": "763fe959-7afc-540e-ae2a-f141a41763f1", 00:14:04.808 "is_configured": true, 00:14:04.808 "data_offset": 2048, 00:14:04.808 "data_size": 63488 00:14:04.808 }, 00:14:04.808 { 00:14:04.808 "name": null, 00:14:04.808 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:04.808 "is_configured": false, 00:14:04.808 "data_offset": 0, 00:14:04.808 "data_size": 63488 00:14:04.808 }, 00:14:04.808 { 00:14:04.808 "name": "BaseBdev3", 00:14:04.808 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:04.808 "is_configured": true, 00:14:04.808 "data_offset": 2048, 00:14:04.808 "data_size": 63488 00:14:04.808 }, 00:14:04.808 { 00:14:04.808 "name": "BaseBdev4", 00:14:04.808 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:04.808 "is_configured": true, 00:14:04.808 "data_offset": 2048, 00:14:04.808 "data_size": 63488 00:14:04.808 } 00:14:04.808 ] 00:14:04.808 }' 00:14:04.808 16:02:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:04.808 16:02:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.065 [2024-11-20 16:02:03.127745] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:05.065 [2024-11-20 16:02:03.127772] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:05.065 [2024-11-20 16:02:03.127836] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:05.065 [2024-11-20 16:02:03.127899] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:05.065 [2024-11-20 16:02:03.127911] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:05.065 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:14:05.322 /dev/nbd0 00:14:05.322 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:05.323 1+0 records in 00:14:05.323 1+0 records out 00:14:05.323 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000193191 s, 21.2 MB/s 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:05.323 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:14:05.580 /dev/nbd1 00:14:05.580 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:05.580 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:05.580 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:14:05.580 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:14:05.580 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:05.580 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:05.580 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:14:05.580 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:14:05.580 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:05.580 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:05.581 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:05.581 1+0 records in 00:14:05.581 1+0 records out 00:14:05.581 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00020968 s, 19.5 MB/s 00:14:05.581 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:05.581 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:14:05.581 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:05.581 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:05.581 16:02:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:14:05.581 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:05.581 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:05.581 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:14:05.581 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:14:05.581 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:05.581 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:05.581 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:05.581 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:14:05.581 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:05.581 16:02:03 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:05.878 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:05.878 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:05.878 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:05.878 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:05.878 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:05.878 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:05.878 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:05.878 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:05.878 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:05.878 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.171 [2024-11-20 16:02:04.245000] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:06.171 [2024-11-20 16:02:04.245045] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:06.171 [2024-11-20 16:02:04.245064] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:14:06.171 [2024-11-20 16:02:04.245072] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:06.171 [2024-11-20 16:02:04.246943] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:06.171 [2024-11-20 16:02:04.246977] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:06.171 [2024-11-20 16:02:04.247049] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:06.171 [2024-11-20 16:02:04.247089] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:06.171 [2024-11-20 16:02:04.247194] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:06.171 [2024-11-20 16:02:04.247269] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:06.171 spare 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.171 [2024-11-20 16:02:04.347350] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:14:06.171 [2024-11-20 16:02:04.347385] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:14:06.171 [2024-11-20 16:02:04.347693] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1c80 00:14:06.171 [2024-11-20 16:02:04.347858] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:14:06.171 [2024-11-20 16:02:04.347874] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:14:06.171 [2024-11-20 16:02:04.348016] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.171 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:06.171 "name": "raid_bdev1", 00:14:06.171 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:06.171 "strip_size_kb": 0, 00:14:06.171 "state": "online", 00:14:06.171 "raid_level": "raid1", 00:14:06.171 "superblock": true, 00:14:06.171 "num_base_bdevs": 4, 00:14:06.171 "num_base_bdevs_discovered": 3, 00:14:06.171 "num_base_bdevs_operational": 3, 00:14:06.171 "base_bdevs_list": [ 00:14:06.171 { 00:14:06.171 "name": "spare", 00:14:06.171 "uuid": "763fe959-7afc-540e-ae2a-f141a41763f1", 00:14:06.171 "is_configured": true, 00:14:06.171 "data_offset": 2048, 00:14:06.171 "data_size": 63488 00:14:06.171 }, 00:14:06.171 { 00:14:06.171 "name": null, 00:14:06.172 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:06.172 "is_configured": false, 00:14:06.172 "data_offset": 2048, 00:14:06.172 "data_size": 63488 00:14:06.172 }, 00:14:06.172 { 00:14:06.172 "name": "BaseBdev3", 00:14:06.172 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:06.172 "is_configured": true, 00:14:06.172 "data_offset": 2048, 00:14:06.172 "data_size": 63488 00:14:06.172 }, 00:14:06.172 { 00:14:06.172 "name": "BaseBdev4", 00:14:06.172 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:06.172 "is_configured": true, 00:14:06.172 "data_offset": 2048, 00:14:06.172 "data_size": 63488 00:14:06.172 } 00:14:06.172 ] 00:14:06.172 }' 00:14:06.172 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:06.172 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:06.737 "name": "raid_bdev1", 00:14:06.737 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:06.737 "strip_size_kb": 0, 00:14:06.737 "state": "online", 00:14:06.737 "raid_level": "raid1", 00:14:06.737 "superblock": true, 00:14:06.737 "num_base_bdevs": 4, 00:14:06.737 "num_base_bdevs_discovered": 3, 00:14:06.737 "num_base_bdevs_operational": 3, 00:14:06.737 "base_bdevs_list": [ 00:14:06.737 { 00:14:06.737 "name": "spare", 00:14:06.737 "uuid": "763fe959-7afc-540e-ae2a-f141a41763f1", 00:14:06.737 "is_configured": true, 00:14:06.737 "data_offset": 2048, 00:14:06.737 "data_size": 63488 00:14:06.737 }, 00:14:06.737 { 00:14:06.737 "name": null, 00:14:06.737 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:06.737 "is_configured": false, 00:14:06.737 "data_offset": 2048, 00:14:06.737 "data_size": 63488 00:14:06.737 }, 00:14:06.737 { 00:14:06.737 "name": "BaseBdev3", 00:14:06.737 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:06.737 "is_configured": true, 00:14:06.737 "data_offset": 2048, 00:14:06.737 "data_size": 63488 00:14:06.737 }, 00:14:06.737 { 00:14:06.737 "name": "BaseBdev4", 00:14:06.737 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:06.737 "is_configured": true, 00:14:06.737 "data_offset": 2048, 00:14:06.737 "data_size": 63488 00:14:06.737 } 00:14:06.737 ] 00:14:06.737 }' 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.737 [2024-11-20 16:02:04.817165] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.737 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:06.737 "name": "raid_bdev1", 00:14:06.737 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:06.737 "strip_size_kb": 0, 00:14:06.737 "state": "online", 00:14:06.737 "raid_level": "raid1", 00:14:06.737 "superblock": true, 00:14:06.737 "num_base_bdevs": 4, 00:14:06.737 "num_base_bdevs_discovered": 2, 00:14:06.737 "num_base_bdevs_operational": 2, 00:14:06.737 "base_bdevs_list": [ 00:14:06.737 { 00:14:06.737 "name": null, 00:14:06.737 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:06.737 "is_configured": false, 00:14:06.737 "data_offset": 0, 00:14:06.737 "data_size": 63488 00:14:06.737 }, 00:14:06.738 { 00:14:06.738 "name": null, 00:14:06.738 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:06.738 "is_configured": false, 00:14:06.738 "data_offset": 2048, 00:14:06.738 "data_size": 63488 00:14:06.738 }, 00:14:06.738 { 00:14:06.738 "name": "BaseBdev3", 00:14:06.738 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:06.738 "is_configured": true, 00:14:06.738 "data_offset": 2048, 00:14:06.738 "data_size": 63488 00:14:06.738 }, 00:14:06.738 { 00:14:06.738 "name": "BaseBdev4", 00:14:06.738 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:06.738 "is_configured": true, 00:14:06.738 "data_offset": 2048, 00:14:06.738 "data_size": 63488 00:14:06.738 } 00:14:06.738 ] 00:14:06.738 }' 00:14:06.738 16:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:06.738 16:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.995 16:02:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:06.995 16:02:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.995 16:02:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.995 [2024-11-20 16:02:05.125226] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:06.995 [2024-11-20 16:02:05.125380] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:14:06.995 [2024-11-20 16:02:05.125392] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:06.995 [2024-11-20 16:02:05.125426] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:06.995 [2024-11-20 16:02:05.132983] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1d50 00:14:06.995 16:02:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.995 16:02:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:14:06.995 [2024-11-20 16:02:05.134569] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:07.928 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:07.928 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:07.928 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:07.928 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:07.928 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:07.928 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.928 16:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.928 16:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:07.928 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:07.928 16:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:08.185 "name": "raid_bdev1", 00:14:08.185 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:08.185 "strip_size_kb": 0, 00:14:08.185 "state": "online", 00:14:08.185 "raid_level": "raid1", 00:14:08.185 "superblock": true, 00:14:08.185 "num_base_bdevs": 4, 00:14:08.185 "num_base_bdevs_discovered": 3, 00:14:08.185 "num_base_bdevs_operational": 3, 00:14:08.185 "process": { 00:14:08.185 "type": "rebuild", 00:14:08.185 "target": "spare", 00:14:08.185 "progress": { 00:14:08.185 "blocks": 20480, 00:14:08.185 "percent": 32 00:14:08.185 } 00:14:08.185 }, 00:14:08.185 "base_bdevs_list": [ 00:14:08.185 { 00:14:08.185 "name": "spare", 00:14:08.185 "uuid": "763fe959-7afc-540e-ae2a-f141a41763f1", 00:14:08.185 "is_configured": true, 00:14:08.185 "data_offset": 2048, 00:14:08.185 "data_size": 63488 00:14:08.185 }, 00:14:08.185 { 00:14:08.185 "name": null, 00:14:08.185 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:08.185 "is_configured": false, 00:14:08.185 "data_offset": 2048, 00:14:08.185 "data_size": 63488 00:14:08.185 }, 00:14:08.185 { 00:14:08.185 "name": "BaseBdev3", 00:14:08.185 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:08.185 "is_configured": true, 00:14:08.185 "data_offset": 2048, 00:14:08.185 "data_size": 63488 00:14:08.185 }, 00:14:08.185 { 00:14:08.185 "name": "BaseBdev4", 00:14:08.185 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:08.185 "is_configured": true, 00:14:08.185 "data_offset": 2048, 00:14:08.185 "data_size": 63488 00:14:08.185 } 00:14:08.185 ] 00:14:08.185 }' 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:08.185 [2024-11-20 16:02:06.244809] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:08.185 [2024-11-20 16:02:06.339859] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:08.185 [2024-11-20 16:02:06.339920] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:08.185 [2024-11-20 16:02:06.339935] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:08.185 [2024-11-20 16:02:06.339941] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:08.185 "name": "raid_bdev1", 00:14:08.185 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:08.185 "strip_size_kb": 0, 00:14:08.185 "state": "online", 00:14:08.185 "raid_level": "raid1", 00:14:08.185 "superblock": true, 00:14:08.185 "num_base_bdevs": 4, 00:14:08.185 "num_base_bdevs_discovered": 2, 00:14:08.185 "num_base_bdevs_operational": 2, 00:14:08.185 "base_bdevs_list": [ 00:14:08.185 { 00:14:08.185 "name": null, 00:14:08.185 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:08.185 "is_configured": false, 00:14:08.185 "data_offset": 0, 00:14:08.185 "data_size": 63488 00:14:08.185 }, 00:14:08.185 { 00:14:08.185 "name": null, 00:14:08.185 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:08.185 "is_configured": false, 00:14:08.185 "data_offset": 2048, 00:14:08.185 "data_size": 63488 00:14:08.185 }, 00:14:08.185 { 00:14:08.185 "name": "BaseBdev3", 00:14:08.185 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:08.185 "is_configured": true, 00:14:08.185 "data_offset": 2048, 00:14:08.185 "data_size": 63488 00:14:08.185 }, 00:14:08.185 { 00:14:08.185 "name": "BaseBdev4", 00:14:08.185 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:08.185 "is_configured": true, 00:14:08.185 "data_offset": 2048, 00:14:08.185 "data_size": 63488 00:14:08.185 } 00:14:08.185 ] 00:14:08.185 }' 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:08.185 16:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:08.442 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:08.442 16:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.442 16:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:08.442 [2024-11-20 16:02:06.680237] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:08.442 [2024-11-20 16:02:06.680295] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:08.442 [2024-11-20 16:02:06.680319] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c380 00:14:08.442 [2024-11-20 16:02:06.680328] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:08.442 [2024-11-20 16:02:06.680711] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:08.442 [2024-11-20 16:02:06.680725] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:08.442 [2024-11-20 16:02:06.680779] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:08.442 [2024-11-20 16:02:06.680788] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:14:08.442 [2024-11-20 16:02:06.680799] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:08.442 [2024-11-20 16:02:06.680817] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:08.442 [2024-11-20 16:02:06.688424] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1e20 00:14:08.442 spare 00:14:08.442 16:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.442 16:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:14:08.442 [2024-11-20 16:02:06.690006] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:09.813 "name": "raid_bdev1", 00:14:09.813 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:09.813 "strip_size_kb": 0, 00:14:09.813 "state": "online", 00:14:09.813 "raid_level": "raid1", 00:14:09.813 "superblock": true, 00:14:09.813 "num_base_bdevs": 4, 00:14:09.813 "num_base_bdevs_discovered": 3, 00:14:09.813 "num_base_bdevs_operational": 3, 00:14:09.813 "process": { 00:14:09.813 "type": "rebuild", 00:14:09.813 "target": "spare", 00:14:09.813 "progress": { 00:14:09.813 "blocks": 20480, 00:14:09.813 "percent": 32 00:14:09.813 } 00:14:09.813 }, 00:14:09.813 "base_bdevs_list": [ 00:14:09.813 { 00:14:09.813 "name": "spare", 00:14:09.813 "uuid": "763fe959-7afc-540e-ae2a-f141a41763f1", 00:14:09.813 "is_configured": true, 00:14:09.813 "data_offset": 2048, 00:14:09.813 "data_size": 63488 00:14:09.813 }, 00:14:09.813 { 00:14:09.813 "name": null, 00:14:09.813 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.813 "is_configured": false, 00:14:09.813 "data_offset": 2048, 00:14:09.813 "data_size": 63488 00:14:09.813 }, 00:14:09.813 { 00:14:09.813 "name": "BaseBdev3", 00:14:09.813 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:09.813 "is_configured": true, 00:14:09.813 "data_offset": 2048, 00:14:09.813 "data_size": 63488 00:14:09.813 }, 00:14:09.813 { 00:14:09.813 "name": "BaseBdev4", 00:14:09.813 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:09.813 "is_configured": true, 00:14:09.813 "data_offset": 2048, 00:14:09.813 "data_size": 63488 00:14:09.813 } 00:14:09.813 ] 00:14:09.813 }' 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:09.813 [2024-11-20 16:02:07.796277] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:09.813 [2024-11-20 16:02:07.895380] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:09.813 [2024-11-20 16:02:07.895448] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:09.813 [2024-11-20 16:02:07.895461] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:09.813 [2024-11-20 16:02:07.895469] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.813 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:09.813 "name": "raid_bdev1", 00:14:09.813 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:09.813 "strip_size_kb": 0, 00:14:09.813 "state": "online", 00:14:09.813 "raid_level": "raid1", 00:14:09.813 "superblock": true, 00:14:09.813 "num_base_bdevs": 4, 00:14:09.813 "num_base_bdevs_discovered": 2, 00:14:09.813 "num_base_bdevs_operational": 2, 00:14:09.813 "base_bdevs_list": [ 00:14:09.813 { 00:14:09.813 "name": null, 00:14:09.813 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.813 "is_configured": false, 00:14:09.813 "data_offset": 0, 00:14:09.813 "data_size": 63488 00:14:09.813 }, 00:14:09.813 { 00:14:09.813 "name": null, 00:14:09.813 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.813 "is_configured": false, 00:14:09.813 "data_offset": 2048, 00:14:09.813 "data_size": 63488 00:14:09.813 }, 00:14:09.813 { 00:14:09.813 "name": "BaseBdev3", 00:14:09.813 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:09.814 "is_configured": true, 00:14:09.814 "data_offset": 2048, 00:14:09.814 "data_size": 63488 00:14:09.814 }, 00:14:09.814 { 00:14:09.814 "name": "BaseBdev4", 00:14:09.814 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:09.814 "is_configured": true, 00:14:09.814 "data_offset": 2048, 00:14:09.814 "data_size": 63488 00:14:09.814 } 00:14:09.814 ] 00:14:09.814 }' 00:14:09.814 16:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:09.814 16:02:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:10.071 "name": "raid_bdev1", 00:14:10.071 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:10.071 "strip_size_kb": 0, 00:14:10.071 "state": "online", 00:14:10.071 "raid_level": "raid1", 00:14:10.071 "superblock": true, 00:14:10.071 "num_base_bdevs": 4, 00:14:10.071 "num_base_bdevs_discovered": 2, 00:14:10.071 "num_base_bdevs_operational": 2, 00:14:10.071 "base_bdevs_list": [ 00:14:10.071 { 00:14:10.071 "name": null, 00:14:10.071 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.071 "is_configured": false, 00:14:10.071 "data_offset": 0, 00:14:10.071 "data_size": 63488 00:14:10.071 }, 00:14:10.071 { 00:14:10.071 "name": null, 00:14:10.071 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.071 "is_configured": false, 00:14:10.071 "data_offset": 2048, 00:14:10.071 "data_size": 63488 00:14:10.071 }, 00:14:10.071 { 00:14:10.071 "name": "BaseBdev3", 00:14:10.071 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:10.071 "is_configured": true, 00:14:10.071 "data_offset": 2048, 00:14:10.071 "data_size": 63488 00:14:10.071 }, 00:14:10.071 { 00:14:10.071 "name": "BaseBdev4", 00:14:10.071 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:10.071 "is_configured": true, 00:14:10.071 "data_offset": 2048, 00:14:10.071 "data_size": 63488 00:14:10.071 } 00:14:10.071 ] 00:14:10.071 }' 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.071 16:02:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.329 16:02:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.329 16:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:10.329 16:02:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.329 16:02:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.329 [2024-11-20 16:02:08.327586] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:10.329 [2024-11-20 16:02:08.327637] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:10.329 [2024-11-20 16:02:08.327651] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c980 00:14:10.329 [2024-11-20 16:02:08.327660] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:10.329 [2024-11-20 16:02:08.328022] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:10.329 [2024-11-20 16:02:08.328048] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:10.329 [2024-11-20 16:02:08.328107] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:14:10.329 [2024-11-20 16:02:08.328119] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:14:10.329 [2024-11-20 16:02:08.328125] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:10.329 [2024-11-20 16:02:08.328140] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:14:10.329 BaseBdev1 00:14:10.329 16:02:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.329 16:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:11.263 "name": "raid_bdev1", 00:14:11.263 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:11.263 "strip_size_kb": 0, 00:14:11.263 "state": "online", 00:14:11.263 "raid_level": "raid1", 00:14:11.263 "superblock": true, 00:14:11.263 "num_base_bdevs": 4, 00:14:11.263 "num_base_bdevs_discovered": 2, 00:14:11.263 "num_base_bdevs_operational": 2, 00:14:11.263 "base_bdevs_list": [ 00:14:11.263 { 00:14:11.263 "name": null, 00:14:11.263 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:11.263 "is_configured": false, 00:14:11.263 "data_offset": 0, 00:14:11.263 "data_size": 63488 00:14:11.263 }, 00:14:11.263 { 00:14:11.263 "name": null, 00:14:11.263 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:11.263 "is_configured": false, 00:14:11.263 "data_offset": 2048, 00:14:11.263 "data_size": 63488 00:14:11.263 }, 00:14:11.263 { 00:14:11.263 "name": "BaseBdev3", 00:14:11.263 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:11.263 "is_configured": true, 00:14:11.263 "data_offset": 2048, 00:14:11.263 "data_size": 63488 00:14:11.263 }, 00:14:11.263 { 00:14:11.263 "name": "BaseBdev4", 00:14:11.263 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:11.263 "is_configured": true, 00:14:11.263 "data_offset": 2048, 00:14:11.263 "data_size": 63488 00:14:11.263 } 00:14:11.263 ] 00:14:11.263 }' 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:11.263 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.534 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:11.534 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:11.534 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:11.534 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:11.534 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:11.534 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:11.534 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.534 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.534 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:11.534 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.534 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:11.534 "name": "raid_bdev1", 00:14:11.534 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:11.534 "strip_size_kb": 0, 00:14:11.534 "state": "online", 00:14:11.534 "raid_level": "raid1", 00:14:11.534 "superblock": true, 00:14:11.534 "num_base_bdevs": 4, 00:14:11.534 "num_base_bdevs_discovered": 2, 00:14:11.534 "num_base_bdevs_operational": 2, 00:14:11.534 "base_bdevs_list": [ 00:14:11.534 { 00:14:11.534 "name": null, 00:14:11.534 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:11.534 "is_configured": false, 00:14:11.534 "data_offset": 0, 00:14:11.534 "data_size": 63488 00:14:11.534 }, 00:14:11.534 { 00:14:11.534 "name": null, 00:14:11.534 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:11.534 "is_configured": false, 00:14:11.534 "data_offset": 2048, 00:14:11.534 "data_size": 63488 00:14:11.534 }, 00:14:11.534 { 00:14:11.534 "name": "BaseBdev3", 00:14:11.534 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:11.534 "is_configured": true, 00:14:11.534 "data_offset": 2048, 00:14:11.534 "data_size": 63488 00:14:11.534 }, 00:14:11.534 { 00:14:11.534 "name": "BaseBdev4", 00:14:11.534 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:11.534 "is_configured": true, 00:14:11.534 "data_offset": 2048, 00:14:11.534 "data_size": 63488 00:14:11.534 } 00:14:11.534 ] 00:14:11.534 }' 00:14:11.534 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:11.534 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:11.534 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:11.535 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:11.535 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:11.535 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:14:11.535 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:11.535 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:14:11.535 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:11.535 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:14:11.535 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:11.535 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:11.535 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.535 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.535 [2024-11-20 16:02:09.755888] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:11.535 [2024-11-20 16:02:09.756031] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:14:11.535 [2024-11-20 16:02:09.756042] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:11.535 request: 00:14:11.535 { 00:14:11.535 "base_bdev": "BaseBdev1", 00:14:11.535 "raid_bdev": "raid_bdev1", 00:14:11.535 "method": "bdev_raid_add_base_bdev", 00:14:11.535 "req_id": 1 00:14:11.535 } 00:14:11.535 Got JSON-RPC error response 00:14:11.535 response: 00:14:11.535 { 00:14:11.535 "code": -22, 00:14:11.535 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:14:11.535 } 00:14:11.535 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:14:11.535 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:14:11.535 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:11.535 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:11.535 16:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:11.535 16:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:14:12.906 16:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:12.906 16:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:12.906 16:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:12.906 16:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:12.906 16:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:12.906 16:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:12.906 16:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:12.906 16:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:12.906 16:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:12.906 16:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:12.906 16:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.906 16:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:12.906 16:02:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.906 16:02:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.906 16:02:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.906 16:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:12.906 "name": "raid_bdev1", 00:14:12.906 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:12.906 "strip_size_kb": 0, 00:14:12.906 "state": "online", 00:14:12.906 "raid_level": "raid1", 00:14:12.906 "superblock": true, 00:14:12.906 "num_base_bdevs": 4, 00:14:12.906 "num_base_bdevs_discovered": 2, 00:14:12.906 "num_base_bdevs_operational": 2, 00:14:12.906 "base_bdevs_list": [ 00:14:12.906 { 00:14:12.906 "name": null, 00:14:12.906 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.906 "is_configured": false, 00:14:12.906 "data_offset": 0, 00:14:12.907 "data_size": 63488 00:14:12.907 }, 00:14:12.907 { 00:14:12.907 "name": null, 00:14:12.907 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.907 "is_configured": false, 00:14:12.907 "data_offset": 2048, 00:14:12.907 "data_size": 63488 00:14:12.907 }, 00:14:12.907 { 00:14:12.907 "name": "BaseBdev3", 00:14:12.907 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:12.907 "is_configured": true, 00:14:12.907 "data_offset": 2048, 00:14:12.907 "data_size": 63488 00:14:12.907 }, 00:14:12.907 { 00:14:12.907 "name": "BaseBdev4", 00:14:12.907 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:12.907 "is_configured": true, 00:14:12.907 "data_offset": 2048, 00:14:12.907 "data_size": 63488 00:14:12.907 } 00:14:12.907 ] 00:14:12.907 }' 00:14:12.907 16:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:12.907 16:02:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:12.907 "name": "raid_bdev1", 00:14:12.907 "uuid": "0351d2f8-099d-48f6-8ae9-46dd080f50e6", 00:14:12.907 "strip_size_kb": 0, 00:14:12.907 "state": "online", 00:14:12.907 "raid_level": "raid1", 00:14:12.907 "superblock": true, 00:14:12.907 "num_base_bdevs": 4, 00:14:12.907 "num_base_bdevs_discovered": 2, 00:14:12.907 "num_base_bdevs_operational": 2, 00:14:12.907 "base_bdevs_list": [ 00:14:12.907 { 00:14:12.907 "name": null, 00:14:12.907 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.907 "is_configured": false, 00:14:12.907 "data_offset": 0, 00:14:12.907 "data_size": 63488 00:14:12.907 }, 00:14:12.907 { 00:14:12.907 "name": null, 00:14:12.907 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.907 "is_configured": false, 00:14:12.907 "data_offset": 2048, 00:14:12.907 "data_size": 63488 00:14:12.907 }, 00:14:12.907 { 00:14:12.907 "name": "BaseBdev3", 00:14:12.907 "uuid": "17c886c0-432e-540b-8b7b-d613fa50ccfd", 00:14:12.907 "is_configured": true, 00:14:12.907 "data_offset": 2048, 00:14:12.907 "data_size": 63488 00:14:12.907 }, 00:14:12.907 { 00:14:12.907 "name": "BaseBdev4", 00:14:12.907 "uuid": "04a372c9-9f7d-5f7a-9e41-cffe0bba4691", 00:14:12.907 "is_configured": true, 00:14:12.907 "data_offset": 2048, 00:14:12.907 "data_size": 63488 00:14:12.907 } 00:14:12.907 ] 00:14:12.907 }' 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 75874 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 75874 ']' 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 75874 00:14:12.907 16:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:14:13.164 16:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:13.164 16:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 75874 00:14:13.164 16:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:13.164 killing process with pid 75874 00:14:13.164 16:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:13.164 16:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 75874' 00:14:13.164 16:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 75874 00:14:13.164 Received shutdown signal, test time was about 60.000000 seconds 00:14:13.164 00:14:13.164 Latency(us) 00:14:13.164 [2024-11-20T16:02:11.415Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:13.164 [2024-11-20T16:02:11.415Z] =================================================================================================================== 00:14:13.164 [2024-11-20T16:02:11.415Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:14:13.164 [2024-11-20 16:02:11.172114] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:13.164 16:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 75874 00:14:13.164 [2024-11-20 16:02:11.172229] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:13.164 [2024-11-20 16:02:11.172290] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:13.164 [2024-11-20 16:02:11.172298] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:14:13.164 [2024-11-20 16:02:11.412109] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:13.731 16:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:14:13.731 00:14:13.731 real 0m24.132s 00:14:13.731 user 0m27.583s 00:14:13.731 sys 0m3.422s 00:14:13.731 ************************************ 00:14:13.731 END TEST raid_rebuild_test_sb 00:14:13.731 ************************************ 00:14:13.731 16:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:13.731 16:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.993 16:02:12 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 4 false true true 00:14:13.993 16:02:12 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:14:13.993 16:02:12 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:13.993 16:02:12 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:13.993 ************************************ 00:14:13.993 START TEST raid_rebuild_test_io 00:14:13.993 ************************************ 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 false true true 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=76632 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 76632 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # '[' -z 76632 ']' 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:13.993 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:13.993 16:02:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:13.993 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:13.993 Zero copy mechanism will not be used. 00:14:13.993 [2024-11-20 16:02:12.121616] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:14:13.993 [2024-11-20 16:02:12.121763] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76632 ] 00:14:14.255 [2024-11-20 16:02:12.279204] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:14.255 [2024-11-20 16:02:12.380281] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:14.578 [2024-11-20 16:02:12.526939] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:14.578 [2024-11-20 16:02:12.526983] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:14.853 16:02:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:14.853 16:02:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # return 0 00:14:14.853 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:14.853 16:02:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:14.853 16:02:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.853 16:02:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:14.853 BaseBdev1_malloc 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:14.853 [2024-11-20 16:02:13.008131] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:14.853 [2024-11-20 16:02:13.008193] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:14.853 [2024-11-20 16:02:13.008216] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:14.853 [2024-11-20 16:02:13.008228] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:14.853 [2024-11-20 16:02:13.010332] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:14.853 [2024-11-20 16:02:13.010370] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:14.853 BaseBdev1 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:14.853 BaseBdev2_malloc 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:14.853 [2024-11-20 16:02:13.048020] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:14.853 [2024-11-20 16:02:13.048177] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:14.853 [2024-11-20 16:02:13.048221] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:14.853 [2024-11-20 16:02:13.048336] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:14.853 [2024-11-20 16:02:13.050459] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:14.853 BaseBdev2 00:14:14.853 [2024-11-20 16:02:13.050566] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:14.853 BaseBdev3_malloc 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.853 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:14.853 [2024-11-20 16:02:13.100845] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:14.853 [2024-11-20 16:02:13.101007] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:14.853 [2024-11-20 16:02:13.101053] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:14.853 [2024-11-20 16:02:13.101110] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:15.111 [2024-11-20 16:02:13.103222] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:15.111 [2024-11-20 16:02:13.103358] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:15.111 BaseBdev3 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.111 BaseBdev4_malloc 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.111 [2024-11-20 16:02:13.144974] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:14:15.111 [2024-11-20 16:02:13.145127] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:15.111 [2024-11-20 16:02:13.145166] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:14:15.111 [2024-11-20 16:02:13.145218] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:15.111 [2024-11-20 16:02:13.147339] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:15.111 [2024-11-20 16:02:13.147449] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:14:15.111 BaseBdev4 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.111 spare_malloc 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.111 spare_delay 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.111 [2024-11-20 16:02:13.193246] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:15.111 [2024-11-20 16:02:13.193297] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:15.111 [2024-11-20 16:02:13.193313] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:14:15.111 [2024-11-20 16:02:13.193323] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:15.111 [2024-11-20 16:02:13.195418] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:15.111 [2024-11-20 16:02:13.195453] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:15.111 spare 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.111 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.111 [2024-11-20 16:02:13.201292] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:15.111 [2024-11-20 16:02:13.203189] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:15.111 [2024-11-20 16:02:13.203333] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:15.111 [2024-11-20 16:02:13.203409] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:15.111 [2024-11-20 16:02:13.203568] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:14:15.111 [2024-11-20 16:02:13.203633] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:14:15.111 [2024-11-20 16:02:13.203927] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:14:15.111 [2024-11-20 16:02:13.204144] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:14:15.112 [2024-11-20 16:02:13.204159] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:14:15.112 [2024-11-20 16:02:13.204301] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:15.112 "name": "raid_bdev1", 00:14:15.112 "uuid": "1369fd3e-2b6c-4364-8d38-27064f0eb483", 00:14:15.112 "strip_size_kb": 0, 00:14:15.112 "state": "online", 00:14:15.112 "raid_level": "raid1", 00:14:15.112 "superblock": false, 00:14:15.112 "num_base_bdevs": 4, 00:14:15.112 "num_base_bdevs_discovered": 4, 00:14:15.112 "num_base_bdevs_operational": 4, 00:14:15.112 "base_bdevs_list": [ 00:14:15.112 { 00:14:15.112 "name": "BaseBdev1", 00:14:15.112 "uuid": "8671c79e-ed25-5824-bed3-dac7c57e1bf3", 00:14:15.112 "is_configured": true, 00:14:15.112 "data_offset": 0, 00:14:15.112 "data_size": 65536 00:14:15.112 }, 00:14:15.112 { 00:14:15.112 "name": "BaseBdev2", 00:14:15.112 "uuid": "11be8003-f23f-5693-b620-870161f5337c", 00:14:15.112 "is_configured": true, 00:14:15.112 "data_offset": 0, 00:14:15.112 "data_size": 65536 00:14:15.112 }, 00:14:15.112 { 00:14:15.112 "name": "BaseBdev3", 00:14:15.112 "uuid": "9c04ec7a-8e1d-5561-bd6d-ec7a70d7c6a4", 00:14:15.112 "is_configured": true, 00:14:15.112 "data_offset": 0, 00:14:15.112 "data_size": 65536 00:14:15.112 }, 00:14:15.112 { 00:14:15.112 "name": "BaseBdev4", 00:14:15.112 "uuid": "19e5c86f-b389-5c3b-947d-99c7d7d60e9b", 00:14:15.112 "is_configured": true, 00:14:15.112 "data_offset": 0, 00:14:15.112 "data_size": 65536 00:14:15.112 } 00:14:15.112 ] 00:14:15.112 }' 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:15.112 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:15.371 [2024-11-20 16:02:13.521745] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.371 [2024-11-20 16:02:13.585369] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:15.371 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:15.372 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:15.372 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:15.372 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.372 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.372 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.372 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:15.372 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.631 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:15.631 "name": "raid_bdev1", 00:14:15.631 "uuid": "1369fd3e-2b6c-4364-8d38-27064f0eb483", 00:14:15.631 "strip_size_kb": 0, 00:14:15.631 "state": "online", 00:14:15.631 "raid_level": "raid1", 00:14:15.631 "superblock": false, 00:14:15.631 "num_base_bdevs": 4, 00:14:15.631 "num_base_bdevs_discovered": 3, 00:14:15.631 "num_base_bdevs_operational": 3, 00:14:15.631 "base_bdevs_list": [ 00:14:15.631 { 00:14:15.631 "name": null, 00:14:15.631 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:15.631 "is_configured": false, 00:14:15.631 "data_offset": 0, 00:14:15.631 "data_size": 65536 00:14:15.631 }, 00:14:15.631 { 00:14:15.631 "name": "BaseBdev2", 00:14:15.631 "uuid": "11be8003-f23f-5693-b620-870161f5337c", 00:14:15.631 "is_configured": true, 00:14:15.631 "data_offset": 0, 00:14:15.631 "data_size": 65536 00:14:15.631 }, 00:14:15.631 { 00:14:15.631 "name": "BaseBdev3", 00:14:15.631 "uuid": "9c04ec7a-8e1d-5561-bd6d-ec7a70d7c6a4", 00:14:15.631 "is_configured": true, 00:14:15.631 "data_offset": 0, 00:14:15.631 "data_size": 65536 00:14:15.631 }, 00:14:15.631 { 00:14:15.631 "name": "BaseBdev4", 00:14:15.631 "uuid": "19e5c86f-b389-5c3b-947d-99c7d7d60e9b", 00:14:15.631 "is_configured": true, 00:14:15.631 "data_offset": 0, 00:14:15.631 "data_size": 65536 00:14:15.631 } 00:14:15.631 ] 00:14:15.631 }' 00:14:15.631 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:15.631 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.631 [2024-11-20 16:02:13.666841] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:14:15.631 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:15.631 Zero copy mechanism will not be used. 00:14:15.631 Running I/O for 60 seconds... 00:14:15.891 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:15.891 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.891 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.891 [2024-11-20 16:02:13.909452] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:15.891 16:02:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.891 16:02:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:15.891 [2024-11-20 16:02:13.971030] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:14:15.891 [2024-11-20 16:02:13.973069] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:15.891 [2024-11-20 16:02:14.106408] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:14:15.891 [2024-11-20 16:02:14.106868] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:14:16.151 [2024-11-20 16:02:14.233686] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:14:16.151 [2024-11-20 16:02:14.234431] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:14:16.411 [2024-11-20 16:02:14.580900] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:14:16.411 [2024-11-20 16:02:14.581575] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:14:16.670 142.00 IOPS, 426.00 MiB/s [2024-11-20T16:02:14.921Z] [2024-11-20 16:02:14.684949] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:14:16.670 [2024-11-20 16:02:14.685182] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:14:16.670 [2024-11-20 16:02:14.904789] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:14:16.670 [2024-11-20 16:02:14.905910] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:14:16.930 16:02:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:16.930 16:02:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:16.930 16:02:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:16.930 16:02:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:16.930 16:02:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:16.930 16:02:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:16.930 16:02:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:16.930 16:02:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:16.930 16:02:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:16.930 16:02:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:16.930 16:02:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:16.930 "name": "raid_bdev1", 00:14:16.930 "uuid": "1369fd3e-2b6c-4364-8d38-27064f0eb483", 00:14:16.930 "strip_size_kb": 0, 00:14:16.930 "state": "online", 00:14:16.930 "raid_level": "raid1", 00:14:16.930 "superblock": false, 00:14:16.930 "num_base_bdevs": 4, 00:14:16.930 "num_base_bdevs_discovered": 4, 00:14:16.930 "num_base_bdevs_operational": 4, 00:14:16.930 "process": { 00:14:16.930 "type": "rebuild", 00:14:16.930 "target": "spare", 00:14:16.930 "progress": { 00:14:16.930 "blocks": 14336, 00:14:16.930 "percent": 21 00:14:16.930 } 00:14:16.930 }, 00:14:16.930 "base_bdevs_list": [ 00:14:16.930 { 00:14:16.930 "name": "spare", 00:14:16.930 "uuid": "7f47406f-932a-5f47-8423-0390181e4a09", 00:14:16.930 "is_configured": true, 00:14:16.930 "data_offset": 0, 00:14:16.930 "data_size": 65536 00:14:16.930 }, 00:14:16.930 { 00:14:16.930 "name": "BaseBdev2", 00:14:16.930 "uuid": "11be8003-f23f-5693-b620-870161f5337c", 00:14:16.930 "is_configured": true, 00:14:16.930 "data_offset": 0, 00:14:16.930 "data_size": 65536 00:14:16.930 }, 00:14:16.930 { 00:14:16.930 "name": "BaseBdev3", 00:14:16.930 "uuid": "9c04ec7a-8e1d-5561-bd6d-ec7a70d7c6a4", 00:14:16.930 "is_configured": true, 00:14:16.930 "data_offset": 0, 00:14:16.930 "data_size": 65536 00:14:16.930 }, 00:14:16.930 { 00:14:16.930 "name": "BaseBdev4", 00:14:16.930 "uuid": "19e5c86f-b389-5c3b-947d-99c7d7d60e9b", 00:14:16.930 "is_configured": true, 00:14:16.930 "data_offset": 0, 00:14:16.930 "data_size": 65536 00:14:16.930 } 00:14:16.930 ] 00:14:16.930 }' 00:14:16.930 16:02:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:16.930 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:16.930 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:16.930 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:16.930 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:16.930 16:02:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:16.930 16:02:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:16.930 [2024-11-20 16:02:15.049657] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:16.930 [2024-11-20 16:02:15.141385] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:14:16.930 [2024-11-20 16:02:15.141751] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:14:16.930 [2024-11-20 16:02:15.142945] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:16.930 [2024-11-20 16:02:15.152504] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:16.930 [2024-11-20 16:02:15.152621] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:16.930 [2024-11-20 16:02:15.152658] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:16.930 [2024-11-20 16:02:15.177655] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006220 00:14:17.189 16:02:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.189 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:14:17.189 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:17.189 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:17.189 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:17.190 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:17.190 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:17.190 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:17.190 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:17.190 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:17.190 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:17.190 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:17.190 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:17.190 16:02:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.190 16:02:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:17.190 16:02:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.190 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:17.190 "name": "raid_bdev1", 00:14:17.190 "uuid": "1369fd3e-2b6c-4364-8d38-27064f0eb483", 00:14:17.190 "strip_size_kb": 0, 00:14:17.190 "state": "online", 00:14:17.190 "raid_level": "raid1", 00:14:17.190 "superblock": false, 00:14:17.190 "num_base_bdevs": 4, 00:14:17.190 "num_base_bdevs_discovered": 3, 00:14:17.190 "num_base_bdevs_operational": 3, 00:14:17.190 "base_bdevs_list": [ 00:14:17.190 { 00:14:17.190 "name": null, 00:14:17.190 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:17.190 "is_configured": false, 00:14:17.190 "data_offset": 0, 00:14:17.190 "data_size": 65536 00:14:17.190 }, 00:14:17.190 { 00:14:17.190 "name": "BaseBdev2", 00:14:17.190 "uuid": "11be8003-f23f-5693-b620-870161f5337c", 00:14:17.190 "is_configured": true, 00:14:17.190 "data_offset": 0, 00:14:17.190 "data_size": 65536 00:14:17.190 }, 00:14:17.190 { 00:14:17.190 "name": "BaseBdev3", 00:14:17.190 "uuid": "9c04ec7a-8e1d-5561-bd6d-ec7a70d7c6a4", 00:14:17.190 "is_configured": true, 00:14:17.190 "data_offset": 0, 00:14:17.190 "data_size": 65536 00:14:17.190 }, 00:14:17.190 { 00:14:17.190 "name": "BaseBdev4", 00:14:17.190 "uuid": "19e5c86f-b389-5c3b-947d-99c7d7d60e9b", 00:14:17.190 "is_configured": true, 00:14:17.190 "data_offset": 0, 00:14:17.190 "data_size": 65536 00:14:17.190 } 00:14:17.190 ] 00:14:17.190 }' 00:14:17.190 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:17.190 16:02:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:17.449 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:17.450 "name": "raid_bdev1", 00:14:17.450 "uuid": "1369fd3e-2b6c-4364-8d38-27064f0eb483", 00:14:17.450 "strip_size_kb": 0, 00:14:17.450 "state": "online", 00:14:17.450 "raid_level": "raid1", 00:14:17.450 "superblock": false, 00:14:17.450 "num_base_bdevs": 4, 00:14:17.450 "num_base_bdevs_discovered": 3, 00:14:17.450 "num_base_bdevs_operational": 3, 00:14:17.450 "base_bdevs_list": [ 00:14:17.450 { 00:14:17.450 "name": null, 00:14:17.450 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:17.450 "is_configured": false, 00:14:17.450 "data_offset": 0, 00:14:17.450 "data_size": 65536 00:14:17.450 }, 00:14:17.450 { 00:14:17.450 "name": "BaseBdev2", 00:14:17.450 "uuid": "11be8003-f23f-5693-b620-870161f5337c", 00:14:17.450 "is_configured": true, 00:14:17.450 "data_offset": 0, 00:14:17.450 "data_size": 65536 00:14:17.450 }, 00:14:17.450 { 00:14:17.450 "name": "BaseBdev3", 00:14:17.450 "uuid": "9c04ec7a-8e1d-5561-bd6d-ec7a70d7c6a4", 00:14:17.450 "is_configured": true, 00:14:17.450 "data_offset": 0, 00:14:17.450 "data_size": 65536 00:14:17.450 }, 00:14:17.450 { 00:14:17.450 "name": "BaseBdev4", 00:14:17.450 "uuid": "19e5c86f-b389-5c3b-947d-99c7d7d60e9b", 00:14:17.450 "is_configured": true, 00:14:17.450 "data_offset": 0, 00:14:17.450 "data_size": 65536 00:14:17.450 } 00:14:17.450 ] 00:14:17.450 }' 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:17.450 [2024-11-20 16:02:15.641061] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.450 16:02:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:17.450 154.50 IOPS, 463.50 MiB/s [2024-11-20T16:02:15.701Z] [2024-11-20 16:02:15.695224] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:14:17.450 [2024-11-20 16:02:15.697211] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:17.710 [2024-11-20 16:02:15.829701] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:14:17.710 [2024-11-20 16:02:15.830217] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:14:17.996 [2024-11-20 16:02:16.056627] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:14:17.996 [2024-11-20 16:02:16.056920] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:14:18.256 [2024-11-20 16:02:16.301659] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:14:18.256 [2024-11-20 16:02:16.420466] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:14:18.517 16:02:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:18.517 16:02:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:18.517 16:02:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:18.517 16:02:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:18.517 16:02:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:18.517 147.00 IOPS, 441.00 MiB/s [2024-11-20T16:02:16.768Z] 16:02:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:18.517 16:02:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:18.517 16:02:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.517 16:02:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:18.517 [2024-11-20 16:02:16.694409] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:14:18.517 16:02:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.517 16:02:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:18.517 "name": "raid_bdev1", 00:14:18.517 "uuid": "1369fd3e-2b6c-4364-8d38-27064f0eb483", 00:14:18.517 "strip_size_kb": 0, 00:14:18.517 "state": "online", 00:14:18.517 "raid_level": "raid1", 00:14:18.517 "superblock": false, 00:14:18.517 "num_base_bdevs": 4, 00:14:18.517 "num_base_bdevs_discovered": 4, 00:14:18.517 "num_base_bdevs_operational": 4, 00:14:18.517 "process": { 00:14:18.517 "type": "rebuild", 00:14:18.517 "target": "spare", 00:14:18.517 "progress": { 00:14:18.517 "blocks": 12288, 00:14:18.517 "percent": 18 00:14:18.517 } 00:14:18.517 }, 00:14:18.517 "base_bdevs_list": [ 00:14:18.517 { 00:14:18.517 "name": "spare", 00:14:18.517 "uuid": "7f47406f-932a-5f47-8423-0390181e4a09", 00:14:18.517 "is_configured": true, 00:14:18.517 "data_offset": 0, 00:14:18.517 "data_size": 65536 00:14:18.517 }, 00:14:18.517 { 00:14:18.517 "name": "BaseBdev2", 00:14:18.517 "uuid": "11be8003-f23f-5693-b620-870161f5337c", 00:14:18.517 "is_configured": true, 00:14:18.517 "data_offset": 0, 00:14:18.517 "data_size": 65536 00:14:18.517 }, 00:14:18.517 { 00:14:18.517 "name": "BaseBdev3", 00:14:18.517 "uuid": "9c04ec7a-8e1d-5561-bd6d-ec7a70d7c6a4", 00:14:18.517 "is_configured": true, 00:14:18.517 "data_offset": 0, 00:14:18.517 "data_size": 65536 00:14:18.517 }, 00:14:18.517 { 00:14:18.517 "name": "BaseBdev4", 00:14:18.517 "uuid": "19e5c86f-b389-5c3b-947d-99c7d7d60e9b", 00:14:18.517 "is_configured": true, 00:14:18.517 "data_offset": 0, 00:14:18.517 "data_size": 65536 00:14:18.517 } 00:14:18.517 ] 00:14:18.517 }' 00:14:18.517 16:02:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:18.517 16:02:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:18.517 16:02:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:18.775 16:02:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:18.775 16:02:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:14:18.775 16:02:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:14:18.775 16:02:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:14:18.775 16:02:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:14:18.775 16:02:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:18.775 16:02:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.776 16:02:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:18.776 [2024-11-20 16:02:16.795794] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:14:18.776 [2024-11-20 16:02:16.805136] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:19.037 [2024-11-20 16:02:17.030585] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000006220 00:14:19.037 [2024-11-20 16:02:17.030627] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d0000063c0 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:19.037 "name": "raid_bdev1", 00:14:19.037 "uuid": "1369fd3e-2b6c-4364-8d38-27064f0eb483", 00:14:19.037 "strip_size_kb": 0, 00:14:19.037 "state": "online", 00:14:19.037 "raid_level": "raid1", 00:14:19.037 "superblock": false, 00:14:19.037 "num_base_bdevs": 4, 00:14:19.037 "num_base_bdevs_discovered": 3, 00:14:19.037 "num_base_bdevs_operational": 3, 00:14:19.037 "process": { 00:14:19.037 "type": "rebuild", 00:14:19.037 "target": "spare", 00:14:19.037 "progress": { 00:14:19.037 "blocks": 18432, 00:14:19.037 "percent": 28 00:14:19.037 } 00:14:19.037 }, 00:14:19.037 "base_bdevs_list": [ 00:14:19.037 { 00:14:19.037 "name": "spare", 00:14:19.037 "uuid": "7f47406f-932a-5f47-8423-0390181e4a09", 00:14:19.037 "is_configured": true, 00:14:19.037 "data_offset": 0, 00:14:19.037 "data_size": 65536 00:14:19.037 }, 00:14:19.037 { 00:14:19.037 "name": null, 00:14:19.037 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:19.037 "is_configured": false, 00:14:19.037 "data_offset": 0, 00:14:19.037 "data_size": 65536 00:14:19.037 }, 00:14:19.037 { 00:14:19.037 "name": "BaseBdev3", 00:14:19.037 "uuid": "9c04ec7a-8e1d-5561-bd6d-ec7a70d7c6a4", 00:14:19.037 "is_configured": true, 00:14:19.037 "data_offset": 0, 00:14:19.037 "data_size": 65536 00:14:19.037 }, 00:14:19.037 { 00:14:19.037 "name": "BaseBdev4", 00:14:19.037 "uuid": "19e5c86f-b389-5c3b-947d-99c7d7d60e9b", 00:14:19.037 "is_configured": true, 00:14:19.037 "data_offset": 0, 00:14:19.037 "data_size": 65536 00:14:19.037 } 00:14:19.037 ] 00:14:19.037 }' 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=391 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:19.037 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:19.037 [2024-11-20 16:02:17.135050] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:14:19.038 [2024-11-20 16:02:17.135442] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:14:19.038 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:19.038 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:19.038 16:02:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.038 16:02:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:19.038 16:02:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.038 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:19.038 "name": "raid_bdev1", 00:14:19.038 "uuid": "1369fd3e-2b6c-4364-8d38-27064f0eb483", 00:14:19.038 "strip_size_kb": 0, 00:14:19.038 "state": "online", 00:14:19.038 "raid_level": "raid1", 00:14:19.038 "superblock": false, 00:14:19.038 "num_base_bdevs": 4, 00:14:19.038 "num_base_bdevs_discovered": 3, 00:14:19.038 "num_base_bdevs_operational": 3, 00:14:19.038 "process": { 00:14:19.038 "type": "rebuild", 00:14:19.038 "target": "spare", 00:14:19.038 "progress": { 00:14:19.038 "blocks": 20480, 00:14:19.038 "percent": 31 00:14:19.038 } 00:14:19.038 }, 00:14:19.038 "base_bdevs_list": [ 00:14:19.038 { 00:14:19.038 "name": "spare", 00:14:19.038 "uuid": "7f47406f-932a-5f47-8423-0390181e4a09", 00:14:19.038 "is_configured": true, 00:14:19.038 "data_offset": 0, 00:14:19.038 "data_size": 65536 00:14:19.038 }, 00:14:19.038 { 00:14:19.038 "name": null, 00:14:19.038 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:19.038 "is_configured": false, 00:14:19.038 "data_offset": 0, 00:14:19.038 "data_size": 65536 00:14:19.038 }, 00:14:19.038 { 00:14:19.038 "name": "BaseBdev3", 00:14:19.038 "uuid": "9c04ec7a-8e1d-5561-bd6d-ec7a70d7c6a4", 00:14:19.038 "is_configured": true, 00:14:19.038 "data_offset": 0, 00:14:19.038 "data_size": 65536 00:14:19.038 }, 00:14:19.038 { 00:14:19.038 "name": "BaseBdev4", 00:14:19.038 "uuid": "19e5c86f-b389-5c3b-947d-99c7d7d60e9b", 00:14:19.038 "is_configured": true, 00:14:19.038 "data_offset": 0, 00:14:19.038 "data_size": 65536 00:14:19.038 } 00:14:19.038 ] 00:14:19.038 }' 00:14:19.038 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:19.038 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:19.038 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:19.038 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:19.038 16:02:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:19.868 123.50 IOPS, 370.50 MiB/s [2024-11-20T16:02:18.119Z] [2024-11-20 16:02:17.940344] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:14:20.129 16:02:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:20.129 16:02:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:20.129 16:02:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:20.129 16:02:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:20.129 16:02:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:20.129 16:02:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:20.129 16:02:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:20.129 16:02:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:20.129 16:02:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.129 16:02:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:20.129 16:02:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.129 16:02:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:20.129 "name": "raid_bdev1", 00:14:20.129 "uuid": "1369fd3e-2b6c-4364-8d38-27064f0eb483", 00:14:20.129 "strip_size_kb": 0, 00:14:20.129 "state": "online", 00:14:20.129 "raid_level": "raid1", 00:14:20.129 "superblock": false, 00:14:20.129 "num_base_bdevs": 4, 00:14:20.129 "num_base_bdevs_discovered": 3, 00:14:20.129 "num_base_bdevs_operational": 3, 00:14:20.129 "process": { 00:14:20.129 "type": "rebuild", 00:14:20.129 "target": "spare", 00:14:20.129 "progress": { 00:14:20.129 "blocks": 34816, 00:14:20.129 "percent": 53 00:14:20.129 } 00:14:20.129 }, 00:14:20.129 "base_bdevs_list": [ 00:14:20.129 { 00:14:20.129 "name": "spare", 00:14:20.129 "uuid": "7f47406f-932a-5f47-8423-0390181e4a09", 00:14:20.129 "is_configured": true, 00:14:20.129 "data_offset": 0, 00:14:20.129 "data_size": 65536 00:14:20.129 }, 00:14:20.129 { 00:14:20.129 "name": null, 00:14:20.129 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:20.129 "is_configured": false, 00:14:20.129 "data_offset": 0, 00:14:20.129 "data_size": 65536 00:14:20.129 }, 00:14:20.129 { 00:14:20.129 "name": "BaseBdev3", 00:14:20.129 "uuid": "9c04ec7a-8e1d-5561-bd6d-ec7a70d7c6a4", 00:14:20.129 "is_configured": true, 00:14:20.129 "data_offset": 0, 00:14:20.129 "data_size": 65536 00:14:20.129 }, 00:14:20.129 { 00:14:20.129 "name": "BaseBdev4", 00:14:20.129 "uuid": "19e5c86f-b389-5c3b-947d-99c7d7d60e9b", 00:14:20.129 "is_configured": true, 00:14:20.129 "data_offset": 0, 00:14:20.129 "data_size": 65536 00:14:20.129 } 00:14:20.129 ] 00:14:20.129 }' 00:14:20.129 16:02:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:20.129 16:02:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:20.129 16:02:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:20.129 16:02:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:20.129 16:02:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:20.390 [2024-11-20 16:02:18.391348] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:14:20.390 [2024-11-20 16:02:18.595389] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:14:21.221 109.40 IOPS, 328.20 MiB/s [2024-11-20T16:02:19.472Z] 16:02:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:21.221 16:02:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:21.221 16:02:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:21.221 16:02:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:21.221 16:02:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:21.221 16:02:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:21.221 16:02:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:21.221 16:02:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.221 16:02:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:21.221 16:02:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:21.221 16:02:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.221 16:02:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:21.221 "name": "raid_bdev1", 00:14:21.221 "uuid": "1369fd3e-2b6c-4364-8d38-27064f0eb483", 00:14:21.221 "strip_size_kb": 0, 00:14:21.221 "state": "online", 00:14:21.221 "raid_level": "raid1", 00:14:21.221 "superblock": false, 00:14:21.221 "num_base_bdevs": 4, 00:14:21.221 "num_base_bdevs_discovered": 3, 00:14:21.221 "num_base_bdevs_operational": 3, 00:14:21.221 "process": { 00:14:21.221 "type": "rebuild", 00:14:21.221 "target": "spare", 00:14:21.221 "progress": { 00:14:21.221 "blocks": 51200, 00:14:21.221 "percent": 78 00:14:21.221 } 00:14:21.221 }, 00:14:21.221 "base_bdevs_list": [ 00:14:21.221 { 00:14:21.221 "name": "spare", 00:14:21.221 "uuid": "7f47406f-932a-5f47-8423-0390181e4a09", 00:14:21.221 "is_configured": true, 00:14:21.221 "data_offset": 0, 00:14:21.221 "data_size": 65536 00:14:21.221 }, 00:14:21.221 { 00:14:21.221 "name": null, 00:14:21.221 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:21.221 "is_configured": false, 00:14:21.221 "data_offset": 0, 00:14:21.221 "data_size": 65536 00:14:21.221 }, 00:14:21.221 { 00:14:21.221 "name": "BaseBdev3", 00:14:21.221 "uuid": "9c04ec7a-8e1d-5561-bd6d-ec7a70d7c6a4", 00:14:21.221 "is_configured": true, 00:14:21.221 "data_offset": 0, 00:14:21.221 "data_size": 65536 00:14:21.221 }, 00:14:21.221 { 00:14:21.221 "name": "BaseBdev4", 00:14:21.221 "uuid": "19e5c86f-b389-5c3b-947d-99c7d7d60e9b", 00:14:21.221 "is_configured": true, 00:14:21.221 "data_offset": 0, 00:14:21.221 "data_size": 65536 00:14:21.221 } 00:14:21.221 ] 00:14:21.221 }' 00:14:21.221 16:02:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:21.221 16:02:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:21.221 16:02:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:21.221 [2024-11-20 16:02:19.412476] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:14:21.221 16:02:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:21.221 16:02:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:21.552 [2024-11-20 16:02:19.653511] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:14:22.123 99.00 IOPS, 297.00 MiB/s [2024-11-20T16:02:20.374Z] [2024-11-20 16:02:20.207617] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:22.123 [2024-11-20 16:02:20.314029] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:22.123 [2024-11-20 16:02:20.316317] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:22.386 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:22.386 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:22.386 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:22.386 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:22.386 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:22.386 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:22.386 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.386 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:22.386 16:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.386 16:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:22.386 16:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.386 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:22.386 "name": "raid_bdev1", 00:14:22.386 "uuid": "1369fd3e-2b6c-4364-8d38-27064f0eb483", 00:14:22.386 "strip_size_kb": 0, 00:14:22.386 "state": "online", 00:14:22.386 "raid_level": "raid1", 00:14:22.386 "superblock": false, 00:14:22.386 "num_base_bdevs": 4, 00:14:22.386 "num_base_bdevs_discovered": 3, 00:14:22.386 "num_base_bdevs_operational": 3, 00:14:22.386 "base_bdevs_list": [ 00:14:22.386 { 00:14:22.386 "name": "spare", 00:14:22.386 "uuid": "7f47406f-932a-5f47-8423-0390181e4a09", 00:14:22.387 "is_configured": true, 00:14:22.387 "data_offset": 0, 00:14:22.387 "data_size": 65536 00:14:22.387 }, 00:14:22.387 { 00:14:22.387 "name": null, 00:14:22.387 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:22.387 "is_configured": false, 00:14:22.387 "data_offset": 0, 00:14:22.387 "data_size": 65536 00:14:22.387 }, 00:14:22.387 { 00:14:22.387 "name": "BaseBdev3", 00:14:22.387 "uuid": "9c04ec7a-8e1d-5561-bd6d-ec7a70d7c6a4", 00:14:22.387 "is_configured": true, 00:14:22.387 "data_offset": 0, 00:14:22.387 "data_size": 65536 00:14:22.387 }, 00:14:22.387 { 00:14:22.387 "name": "BaseBdev4", 00:14:22.387 "uuid": "19e5c86f-b389-5c3b-947d-99c7d7d60e9b", 00:14:22.387 "is_configured": true, 00:14:22.387 "data_offset": 0, 00:14:22.387 "data_size": 65536 00:14:22.387 } 00:14:22.387 ] 00:14:22.387 }' 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:22.387 "name": "raid_bdev1", 00:14:22.387 "uuid": "1369fd3e-2b6c-4364-8d38-27064f0eb483", 00:14:22.387 "strip_size_kb": 0, 00:14:22.387 "state": "online", 00:14:22.387 "raid_level": "raid1", 00:14:22.387 "superblock": false, 00:14:22.387 "num_base_bdevs": 4, 00:14:22.387 "num_base_bdevs_discovered": 3, 00:14:22.387 "num_base_bdevs_operational": 3, 00:14:22.387 "base_bdevs_list": [ 00:14:22.387 { 00:14:22.387 "name": "spare", 00:14:22.387 "uuid": "7f47406f-932a-5f47-8423-0390181e4a09", 00:14:22.387 "is_configured": true, 00:14:22.387 "data_offset": 0, 00:14:22.387 "data_size": 65536 00:14:22.387 }, 00:14:22.387 { 00:14:22.387 "name": null, 00:14:22.387 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:22.387 "is_configured": false, 00:14:22.387 "data_offset": 0, 00:14:22.387 "data_size": 65536 00:14:22.387 }, 00:14:22.387 { 00:14:22.387 "name": "BaseBdev3", 00:14:22.387 "uuid": "9c04ec7a-8e1d-5561-bd6d-ec7a70d7c6a4", 00:14:22.387 "is_configured": true, 00:14:22.387 "data_offset": 0, 00:14:22.387 "data_size": 65536 00:14:22.387 }, 00:14:22.387 { 00:14:22.387 "name": "BaseBdev4", 00:14:22.387 "uuid": "19e5c86f-b389-5c3b-947d-99c7d7d60e9b", 00:14:22.387 "is_configured": true, 00:14:22.387 "data_offset": 0, 00:14:22.387 "data_size": 65536 00:14:22.387 } 00:14:22.387 ] 00:14:22.387 }' 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.387 16:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:22.647 16:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.647 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:22.647 "name": "raid_bdev1", 00:14:22.647 "uuid": "1369fd3e-2b6c-4364-8d38-27064f0eb483", 00:14:22.647 "strip_size_kb": 0, 00:14:22.647 "state": "online", 00:14:22.647 "raid_level": "raid1", 00:14:22.647 "superblock": false, 00:14:22.647 "num_base_bdevs": 4, 00:14:22.647 "num_base_bdevs_discovered": 3, 00:14:22.647 "num_base_bdevs_operational": 3, 00:14:22.647 "base_bdevs_list": [ 00:14:22.647 { 00:14:22.647 "name": "spare", 00:14:22.647 "uuid": "7f47406f-932a-5f47-8423-0390181e4a09", 00:14:22.647 "is_configured": true, 00:14:22.647 "data_offset": 0, 00:14:22.647 "data_size": 65536 00:14:22.647 }, 00:14:22.647 { 00:14:22.647 "name": null, 00:14:22.647 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:22.647 "is_configured": false, 00:14:22.647 "data_offset": 0, 00:14:22.647 "data_size": 65536 00:14:22.647 }, 00:14:22.647 { 00:14:22.647 "name": "BaseBdev3", 00:14:22.647 "uuid": "9c04ec7a-8e1d-5561-bd6d-ec7a70d7c6a4", 00:14:22.647 "is_configured": true, 00:14:22.647 "data_offset": 0, 00:14:22.647 "data_size": 65536 00:14:22.647 }, 00:14:22.647 { 00:14:22.647 "name": "BaseBdev4", 00:14:22.647 "uuid": "19e5c86f-b389-5c3b-947d-99c7d7d60e9b", 00:14:22.647 "is_configured": true, 00:14:22.647 "data_offset": 0, 00:14:22.647 "data_size": 65536 00:14:22.647 } 00:14:22.647 ] 00:14:22.647 }' 00:14:22.647 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:22.647 16:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:22.908 91.29 IOPS, 273.86 MiB/s [2024-11-20T16:02:21.159Z] 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:22.908 16:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.908 16:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:22.908 [2024-11-20 16:02:20.958092] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:22.908 [2024-11-20 16:02:20.958219] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:22.908 00:14:22.908 Latency(us) 00:14:22.908 [2024-11-20T16:02:21.159Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:22.908 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:14:22.908 raid_bdev1 : 7.31 89.20 267.61 0.00 0.00 15121.82 302.47 119376.34 00:14:22.908 [2024-11-20T16:02:21.159Z] =================================================================================================================== 00:14:22.908 [2024-11-20T16:02:21.159Z] Total : 89.20 267.61 0.00 0.00 15121.82 302.47 119376.34 00:14:22.908 [2024-11-20 16:02:20.993448] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:22.908 [2024-11-20 16:02:20.993511] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:22.908 [2024-11-20 16:02:20.993613] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:22.909 [2024-11-20 16:02:20.993623] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:14:22.909 { 00:14:22.909 "results": [ 00:14:22.909 { 00:14:22.909 "job": "raid_bdev1", 00:14:22.909 "core_mask": "0x1", 00:14:22.909 "workload": "randrw", 00:14:22.909 "percentage": 50, 00:14:22.909 "status": "finished", 00:14:22.909 "queue_depth": 2, 00:14:22.909 "io_size": 3145728, 00:14:22.909 "runtime": 7.309065, 00:14:22.909 "iops": 89.2042963087618, 00:14:22.909 "mibps": 267.6128889262854, 00:14:22.909 "io_failed": 0, 00:14:22.909 "io_timeout": 0, 00:14:22.909 "avg_latency_us": 15121.818518168946, 00:14:22.909 "min_latency_us": 302.4738461538462, 00:14:22.909 "max_latency_us": 119376.34461538462 00:14:22.909 } 00:14:22.909 ], 00:14:22.909 "core_count": 1 00:14:22.909 } 00:14:22.909 16:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.909 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.909 16:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:14:22.909 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.909 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:22.909 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.909 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:22.909 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:22.909 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:14:22.909 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:14:22.909 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:22.909 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:14:22.909 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:22.909 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:22.909 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:22.909 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:14:22.909 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:22.909 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:22.909 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:14:23.170 /dev/nbd0 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:23.170 1+0 records in 00:14:23.170 1+0 records out 00:14:23.170 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000507652 s, 8.1 MB/s 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@728 -- # continue 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:23.170 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:14:23.431 /dev/nbd1 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:23.431 1+0 records in 00:14:23.431 1+0 records out 00:14:23.431 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000394869 s, 10.4 MB/s 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:14:23.431 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:23.432 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:23.693 16:02:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:14:23.953 /dev/nbd1 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:23.953 1+0 records in 00:14:23.953 1+0 records out 00:14:23.953 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000260118 s, 15.7 MB/s 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:23.953 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:24.213 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:24.213 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:24.213 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:24.213 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:24.213 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:24.213 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:24.213 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:14:24.213 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:14:24.213 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:24.213 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:24.213 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:24.213 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:24.213 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:14:24.213 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:24.213 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 76632 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # '[' -z 76632 ']' 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # kill -0 76632 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # uname 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76632 00:14:24.472 killing process with pid 76632 00:14:24.472 Received shutdown signal, test time was about 8.982494 seconds 00:14:24.472 00:14:24.472 Latency(us) 00:14:24.472 [2024-11-20T16:02:22.723Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:24.472 [2024-11-20T16:02:22.723Z] =================================================================================================================== 00:14:24.472 [2024-11-20T16:02:22.723Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76632' 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@973 -- # kill 76632 00:14:24.472 [2024-11-20 16:02:22.651453] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:24.472 16:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@978 -- # wait 76632 00:14:24.732 [2024-11-20 16:02:22.928052] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:25.675 ************************************ 00:14:25.675 END TEST raid_rebuild_test_io 00:14:25.675 ************************************ 00:14:25.675 16:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:14:25.675 00:14:25.675 real 0m11.682s 00:14:25.675 user 0m14.448s 00:14:25.675 sys 0m1.334s 00:14:25.675 16:02:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:25.675 16:02:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:25.675 16:02:23 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 4 true true true 00:14:25.675 16:02:23 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:14:25.675 16:02:23 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:25.675 16:02:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:25.675 ************************************ 00:14:25.675 START TEST raid_rebuild_test_sb_io 00:14:25.675 ************************************ 00:14:25.675 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 true true true 00:14:25.675 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:14:25.675 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=77026 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 77026 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # '[' -z 77026 ']' 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:25.676 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:25.676 16:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:25.676 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:25.676 Zero copy mechanism will not be used. 00:14:25.676 [2024-11-20 16:02:23.865974] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:14:25.676 [2024-11-20 16:02:23.866093] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77026 ] 00:14:25.938 [2024-11-20 16:02:24.024765] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:25.938 [2024-11-20 16:02:24.126805] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:26.197 [2024-11-20 16:02:24.263626] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:26.197 [2024-11-20 16:02:24.263659] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:26.766 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:26.766 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # return 0 00:14:26.766 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:26.766 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:26.766 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.766 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:26.766 BaseBdev1_malloc 00:14:26.766 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.766 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:26.766 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.766 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:26.766 [2024-11-20 16:02:24.756913] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:26.767 [2024-11-20 16:02:24.756974] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:26.767 [2024-11-20 16:02:24.756996] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:26.767 [2024-11-20 16:02:24.757008] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:26.767 [2024-11-20 16:02:24.759151] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:26.767 [2024-11-20 16:02:24.759298] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:26.767 BaseBdev1 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:26.767 BaseBdev2_malloc 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:26.767 [2024-11-20 16:02:24.793575] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:26.767 [2024-11-20 16:02:24.793639] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:26.767 [2024-11-20 16:02:24.793663] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:26.767 [2024-11-20 16:02:24.793690] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:26.767 [2024-11-20 16:02:24.795907] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:26.767 [2024-11-20 16:02:24.795943] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:26.767 BaseBdev2 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:26.767 BaseBdev3_malloc 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:26.767 [2024-11-20 16:02:24.850016] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:26.767 [2024-11-20 16:02:24.850078] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:26.767 [2024-11-20 16:02:24.850102] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:26.767 [2024-11-20 16:02:24.850112] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:26.767 [2024-11-20 16:02:24.852294] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:26.767 [2024-11-20 16:02:24.852447] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:26.767 BaseBdev3 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:26.767 BaseBdev4_malloc 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:26.767 [2024-11-20 16:02:24.891108] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:14:26.767 [2024-11-20 16:02:24.891280] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:26.767 [2024-11-20 16:02:24.891306] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:14:26.767 [2024-11-20 16:02:24.891317] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:26.767 [2024-11-20 16:02:24.893493] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:26.767 [2024-11-20 16:02:24.893534] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:14:26.767 BaseBdev4 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:26.767 spare_malloc 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:26.767 spare_delay 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:26.767 [2024-11-20 16:02:24.940141] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:26.767 [2024-11-20 16:02:24.940192] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:26.767 [2024-11-20 16:02:24.940211] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:14:26.767 [2024-11-20 16:02:24.940223] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:26.767 [2024-11-20 16:02:24.942444] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:26.767 [2024-11-20 16:02:24.942591] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:26.767 spare 00:14:26.767 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:26.768 [2024-11-20 16:02:24.948185] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:26.768 [2024-11-20 16:02:24.950089] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:26.768 [2024-11-20 16:02:24.950151] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:26.768 [2024-11-20 16:02:24.950203] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:26.768 [2024-11-20 16:02:24.950387] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:14:26.768 [2024-11-20 16:02:24.950400] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:14:26.768 [2024-11-20 16:02:24.950681] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:14:26.768 [2024-11-20 16:02:24.950842] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:14:26.768 [2024-11-20 16:02:24.950851] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:14:26.768 [2024-11-20 16:02:24.951003] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:26.768 "name": "raid_bdev1", 00:14:26.768 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:26.768 "strip_size_kb": 0, 00:14:26.768 "state": "online", 00:14:26.768 "raid_level": "raid1", 00:14:26.768 "superblock": true, 00:14:26.768 "num_base_bdevs": 4, 00:14:26.768 "num_base_bdevs_discovered": 4, 00:14:26.768 "num_base_bdevs_operational": 4, 00:14:26.768 "base_bdevs_list": [ 00:14:26.768 { 00:14:26.768 "name": "BaseBdev1", 00:14:26.768 "uuid": "5ac201db-7570-5172-b24d-2dbce69a4480", 00:14:26.768 "is_configured": true, 00:14:26.768 "data_offset": 2048, 00:14:26.768 "data_size": 63488 00:14:26.768 }, 00:14:26.768 { 00:14:26.768 "name": "BaseBdev2", 00:14:26.768 "uuid": "04ca1b39-a092-58b7-b463-f63d6c637191", 00:14:26.768 "is_configured": true, 00:14:26.768 "data_offset": 2048, 00:14:26.768 "data_size": 63488 00:14:26.768 }, 00:14:26.768 { 00:14:26.768 "name": "BaseBdev3", 00:14:26.768 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:26.768 "is_configured": true, 00:14:26.768 "data_offset": 2048, 00:14:26.768 "data_size": 63488 00:14:26.768 }, 00:14:26.768 { 00:14:26.768 "name": "BaseBdev4", 00:14:26.768 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:26.768 "is_configured": true, 00:14:26.768 "data_offset": 2048, 00:14:26.768 "data_size": 63488 00:14:26.768 } 00:14:26.768 ] 00:14:26.768 }' 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:26.768 16:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:27.028 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:27.028 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:27.028 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.028 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:27.028 [2024-11-20 16:02:25.264610] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:27.287 [2024-11-20 16:02:25.328241] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:27.287 "name": "raid_bdev1", 00:14:27.287 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:27.287 "strip_size_kb": 0, 00:14:27.287 "state": "online", 00:14:27.287 "raid_level": "raid1", 00:14:27.287 "superblock": true, 00:14:27.287 "num_base_bdevs": 4, 00:14:27.287 "num_base_bdevs_discovered": 3, 00:14:27.287 "num_base_bdevs_operational": 3, 00:14:27.287 "base_bdevs_list": [ 00:14:27.287 { 00:14:27.287 "name": null, 00:14:27.287 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:27.287 "is_configured": false, 00:14:27.287 "data_offset": 0, 00:14:27.287 "data_size": 63488 00:14:27.287 }, 00:14:27.287 { 00:14:27.287 "name": "BaseBdev2", 00:14:27.287 "uuid": "04ca1b39-a092-58b7-b463-f63d6c637191", 00:14:27.287 "is_configured": true, 00:14:27.287 "data_offset": 2048, 00:14:27.287 "data_size": 63488 00:14:27.287 }, 00:14:27.287 { 00:14:27.287 "name": "BaseBdev3", 00:14:27.287 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:27.287 "is_configured": true, 00:14:27.287 "data_offset": 2048, 00:14:27.287 "data_size": 63488 00:14:27.287 }, 00:14:27.287 { 00:14:27.287 "name": "BaseBdev4", 00:14:27.287 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:27.287 "is_configured": true, 00:14:27.287 "data_offset": 2048, 00:14:27.287 "data_size": 63488 00:14:27.287 } 00:14:27.287 ] 00:14:27.287 }' 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:27.287 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:27.287 [2024-11-20 16:02:25.413704] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:14:27.287 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:27.287 Zero copy mechanism will not be used. 00:14:27.287 Running I/O for 60 seconds... 00:14:27.549 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:27.549 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.549 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:27.549 [2024-11-20 16:02:25.652296] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:27.549 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.549 16:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:27.549 [2024-11-20 16:02:25.735858] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:14:27.549 [2024-11-20 16:02:25.737915] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:27.810 [2024-11-20 16:02:25.855174] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:14:27.810 [2024-11-20 16:02:25.855597] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:14:28.072 [2024-11-20 16:02:26.082696] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:14:28.072 [2024-11-20 16:02:26.082943] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:14:28.331 [2024-11-20 16:02:26.342752] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:14:28.331 163.00 IOPS, 489.00 MiB/s [2024-11-20T16:02:26.582Z] [2024-11-20 16:02:26.485525] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:14:28.590 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:28.590 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:28.590 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:28.590 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:28.590 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:28.590 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:28.590 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:28.590 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.590 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:28.590 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.590 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:28.590 "name": "raid_bdev1", 00:14:28.590 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:28.590 "strip_size_kb": 0, 00:14:28.590 "state": "online", 00:14:28.590 "raid_level": "raid1", 00:14:28.590 "superblock": true, 00:14:28.590 "num_base_bdevs": 4, 00:14:28.590 "num_base_bdevs_discovered": 4, 00:14:28.590 "num_base_bdevs_operational": 4, 00:14:28.590 "process": { 00:14:28.590 "type": "rebuild", 00:14:28.590 "target": "spare", 00:14:28.590 "progress": { 00:14:28.590 "blocks": 10240, 00:14:28.590 "percent": 16 00:14:28.590 } 00:14:28.590 }, 00:14:28.590 "base_bdevs_list": [ 00:14:28.590 { 00:14:28.590 "name": "spare", 00:14:28.590 "uuid": "9b4e9028-25d6-50e1-8f5c-fcb2012fb089", 00:14:28.590 "is_configured": true, 00:14:28.590 "data_offset": 2048, 00:14:28.590 "data_size": 63488 00:14:28.590 }, 00:14:28.590 { 00:14:28.590 "name": "BaseBdev2", 00:14:28.590 "uuid": "04ca1b39-a092-58b7-b463-f63d6c637191", 00:14:28.590 "is_configured": true, 00:14:28.590 "data_offset": 2048, 00:14:28.590 "data_size": 63488 00:14:28.590 }, 00:14:28.590 { 00:14:28.590 "name": "BaseBdev3", 00:14:28.590 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:28.590 "is_configured": true, 00:14:28.590 "data_offset": 2048, 00:14:28.590 "data_size": 63488 00:14:28.590 }, 00:14:28.590 { 00:14:28.590 "name": "BaseBdev4", 00:14:28.590 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:28.590 "is_configured": true, 00:14:28.590 "data_offset": 2048, 00:14:28.591 "data_size": 63488 00:14:28.591 } 00:14:28.591 ] 00:14:28.591 }' 00:14:28.591 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:28.591 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:28.591 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:28.591 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:28.591 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:28.591 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.591 16:02:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:28.591 [2024-11-20 16:02:26.793480] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:28.851 [2024-11-20 16:02:26.853084] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:14:28.851 [2024-11-20 16:02:26.955202] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:28.851 [2024-11-20 16:02:26.965138] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:28.851 [2024-11-20 16:02:26.965190] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:28.851 [2024-11-20 16:02:26.965203] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:28.851 [2024-11-20 16:02:26.990134] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006220 00:14:28.851 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.851 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:14:28.851 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:28.851 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:28.851 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:28.851 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:28.851 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:28.851 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:28.851 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:28.851 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:28.851 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:28.851 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:28.851 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:28.851 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.851 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:28.851 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.851 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:28.851 "name": "raid_bdev1", 00:14:28.851 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:28.851 "strip_size_kb": 0, 00:14:28.851 "state": "online", 00:14:28.851 "raid_level": "raid1", 00:14:28.851 "superblock": true, 00:14:28.851 "num_base_bdevs": 4, 00:14:28.851 "num_base_bdevs_discovered": 3, 00:14:28.851 "num_base_bdevs_operational": 3, 00:14:28.851 "base_bdevs_list": [ 00:14:28.851 { 00:14:28.851 "name": null, 00:14:28.851 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:28.851 "is_configured": false, 00:14:28.851 "data_offset": 0, 00:14:28.851 "data_size": 63488 00:14:28.851 }, 00:14:28.851 { 00:14:28.851 "name": "BaseBdev2", 00:14:28.851 "uuid": "04ca1b39-a092-58b7-b463-f63d6c637191", 00:14:28.851 "is_configured": true, 00:14:28.851 "data_offset": 2048, 00:14:28.851 "data_size": 63488 00:14:28.851 }, 00:14:28.851 { 00:14:28.851 "name": "BaseBdev3", 00:14:28.851 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:28.851 "is_configured": true, 00:14:28.851 "data_offset": 2048, 00:14:28.851 "data_size": 63488 00:14:28.851 }, 00:14:28.851 { 00:14:28.851 "name": "BaseBdev4", 00:14:28.851 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:28.851 "is_configured": true, 00:14:28.851 "data_offset": 2048, 00:14:28.851 "data_size": 63488 00:14:28.851 } 00:14:28.852 ] 00:14:28.852 }' 00:14:28.852 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:28.852 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:29.112 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:29.112 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:29.112 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:29.112 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:29.112 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:29.112 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.112 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.112 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:29.112 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:29.112 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.374 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:29.374 "name": "raid_bdev1", 00:14:29.374 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:29.374 "strip_size_kb": 0, 00:14:29.374 "state": "online", 00:14:29.374 "raid_level": "raid1", 00:14:29.374 "superblock": true, 00:14:29.374 "num_base_bdevs": 4, 00:14:29.374 "num_base_bdevs_discovered": 3, 00:14:29.374 "num_base_bdevs_operational": 3, 00:14:29.374 "base_bdevs_list": [ 00:14:29.374 { 00:14:29.374 "name": null, 00:14:29.374 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:29.374 "is_configured": false, 00:14:29.374 "data_offset": 0, 00:14:29.374 "data_size": 63488 00:14:29.374 }, 00:14:29.374 { 00:14:29.374 "name": "BaseBdev2", 00:14:29.374 "uuid": "04ca1b39-a092-58b7-b463-f63d6c637191", 00:14:29.374 "is_configured": true, 00:14:29.374 "data_offset": 2048, 00:14:29.374 "data_size": 63488 00:14:29.374 }, 00:14:29.374 { 00:14:29.374 "name": "BaseBdev3", 00:14:29.374 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:29.374 "is_configured": true, 00:14:29.374 "data_offset": 2048, 00:14:29.374 "data_size": 63488 00:14:29.374 }, 00:14:29.374 { 00:14:29.374 "name": "BaseBdev4", 00:14:29.374 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:29.374 "is_configured": true, 00:14:29.374 "data_offset": 2048, 00:14:29.374 "data_size": 63488 00:14:29.374 } 00:14:29.374 ] 00:14:29.374 }' 00:14:29.374 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:29.374 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:29.374 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:29.374 142.00 IOPS, 426.00 MiB/s [2024-11-20T16:02:27.625Z] 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:29.374 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:29.374 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.374 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:29.374 [2024-11-20 16:02:27.447222] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:29.374 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.375 16:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:29.375 [2024-11-20 16:02:27.492175] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:14:29.375 [2024-11-20 16:02:27.494204] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:29.375 [2024-11-20 16:02:27.611722] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:14:29.375 [2024-11-20 16:02:27.612849] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:14:29.635 [2024-11-20 16:02:27.831382] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:14:29.635 [2024-11-20 16:02:27.832038] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:14:30.206 [2024-11-20 16:02:28.165196] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:14:30.206 [2024-11-20 16:02:28.299171] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:14:30.466 126.00 IOPS, 378.00 MiB/s [2024-11-20T16:02:28.717Z] 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:30.466 "name": "raid_bdev1", 00:14:30.466 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:30.466 "strip_size_kb": 0, 00:14:30.466 "state": "online", 00:14:30.466 "raid_level": "raid1", 00:14:30.466 "superblock": true, 00:14:30.466 "num_base_bdevs": 4, 00:14:30.466 "num_base_bdevs_discovered": 4, 00:14:30.466 "num_base_bdevs_operational": 4, 00:14:30.466 "process": { 00:14:30.466 "type": "rebuild", 00:14:30.466 "target": "spare", 00:14:30.466 "progress": { 00:14:30.466 "blocks": 12288, 00:14:30.466 "percent": 19 00:14:30.466 } 00:14:30.466 }, 00:14:30.466 "base_bdevs_list": [ 00:14:30.466 { 00:14:30.466 "name": "spare", 00:14:30.466 "uuid": "9b4e9028-25d6-50e1-8f5c-fcb2012fb089", 00:14:30.466 "is_configured": true, 00:14:30.466 "data_offset": 2048, 00:14:30.466 "data_size": 63488 00:14:30.466 }, 00:14:30.466 { 00:14:30.466 "name": "BaseBdev2", 00:14:30.466 "uuid": "04ca1b39-a092-58b7-b463-f63d6c637191", 00:14:30.466 "is_configured": true, 00:14:30.466 "data_offset": 2048, 00:14:30.466 "data_size": 63488 00:14:30.466 }, 00:14:30.466 { 00:14:30.466 "name": "BaseBdev3", 00:14:30.466 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:30.466 "is_configured": true, 00:14:30.466 "data_offset": 2048, 00:14:30.466 "data_size": 63488 00:14:30.466 }, 00:14:30.466 { 00:14:30.466 "name": "BaseBdev4", 00:14:30.466 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:30.466 "is_configured": true, 00:14:30.466 "data_offset": 2048, 00:14:30.466 "data_size": 63488 00:14:30.466 } 00:14:30.466 ] 00:14:30.466 }' 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:14:30.466 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:30.466 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:30.467 [2024-11-20 16:02:28.577444] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:30.467 [2024-11-20 16:02:28.629364] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:14:30.467 [2024-11-20 16:02:28.630092] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:14:30.750 [2024-11-20 16:02:28.846523] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000006220 00:14:30.750 [2024-11-20 16:02:28.846712] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d0000063c0 00:14:30.750 [2024-11-20 16:02:28.847846] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:14:30.750 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:30.751 "name": "raid_bdev1", 00:14:30.751 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:30.751 "strip_size_kb": 0, 00:14:30.751 "state": "online", 00:14:30.751 "raid_level": "raid1", 00:14:30.751 "superblock": true, 00:14:30.751 "num_base_bdevs": 4, 00:14:30.751 "num_base_bdevs_discovered": 3, 00:14:30.751 "num_base_bdevs_operational": 3, 00:14:30.751 "process": { 00:14:30.751 "type": "rebuild", 00:14:30.751 "target": "spare", 00:14:30.751 "progress": { 00:14:30.751 "blocks": 16384, 00:14:30.751 "percent": 25 00:14:30.751 } 00:14:30.751 }, 00:14:30.751 "base_bdevs_list": [ 00:14:30.751 { 00:14:30.751 "name": "spare", 00:14:30.751 "uuid": "9b4e9028-25d6-50e1-8f5c-fcb2012fb089", 00:14:30.751 "is_configured": true, 00:14:30.751 "data_offset": 2048, 00:14:30.751 "data_size": 63488 00:14:30.751 }, 00:14:30.751 { 00:14:30.751 "name": null, 00:14:30.751 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:30.751 "is_configured": false, 00:14:30.751 "data_offset": 0, 00:14:30.751 "data_size": 63488 00:14:30.751 }, 00:14:30.751 { 00:14:30.751 "name": "BaseBdev3", 00:14:30.751 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:30.751 "is_configured": true, 00:14:30.751 "data_offset": 2048, 00:14:30.751 "data_size": 63488 00:14:30.751 }, 00:14:30.751 { 00:14:30.751 "name": "BaseBdev4", 00:14:30.751 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:30.751 "is_configured": true, 00:14:30.751 "data_offset": 2048, 00:14:30.751 "data_size": 63488 00:14:30.751 } 00:14:30.751 ] 00:14:30.751 }' 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=402 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:30.751 16:02:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:31.011 16:02:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:31.011 "name": "raid_bdev1", 00:14:31.011 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:31.011 "strip_size_kb": 0, 00:14:31.011 "state": "online", 00:14:31.011 "raid_level": "raid1", 00:14:31.011 "superblock": true, 00:14:31.011 "num_base_bdevs": 4, 00:14:31.011 "num_base_bdevs_discovered": 3, 00:14:31.011 "num_base_bdevs_operational": 3, 00:14:31.011 "process": { 00:14:31.011 "type": "rebuild", 00:14:31.011 "target": "spare", 00:14:31.011 "progress": { 00:14:31.011 "blocks": 16384, 00:14:31.011 "percent": 25 00:14:31.011 } 00:14:31.011 }, 00:14:31.011 "base_bdevs_list": [ 00:14:31.011 { 00:14:31.011 "name": "spare", 00:14:31.011 "uuid": "9b4e9028-25d6-50e1-8f5c-fcb2012fb089", 00:14:31.011 "is_configured": true, 00:14:31.011 "data_offset": 2048, 00:14:31.011 "data_size": 63488 00:14:31.011 }, 00:14:31.011 { 00:14:31.011 "name": null, 00:14:31.011 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:31.011 "is_configured": false, 00:14:31.011 "data_offset": 0, 00:14:31.011 "data_size": 63488 00:14:31.011 }, 00:14:31.011 { 00:14:31.011 "name": "BaseBdev3", 00:14:31.011 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:31.011 "is_configured": true, 00:14:31.011 "data_offset": 2048, 00:14:31.011 "data_size": 63488 00:14:31.011 }, 00:14:31.011 { 00:14:31.011 "name": "BaseBdev4", 00:14:31.011 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:31.011 "is_configured": true, 00:14:31.011 "data_offset": 2048, 00:14:31.011 "data_size": 63488 00:14:31.011 } 00:14:31.011 ] 00:14:31.011 }' 00:14:31.011 16:02:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:31.011 16:02:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:31.011 16:02:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:31.011 16:02:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:31.011 16:02:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:31.011 [2024-11-20 16:02:29.161386] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:14:31.011 [2024-11-20 16:02:29.161929] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:14:31.270 [2024-11-20 16:02:29.365373] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:14:31.841 117.00 IOPS, 351.00 MiB/s [2024-11-20T16:02:30.092Z] 16:02:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:31.841 16:02:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:31.841 16:02:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:31.841 16:02:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:31.841 16:02:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:31.841 16:02:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:31.841 16:02:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:31.841 16:02:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:31.841 16:02:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:31.841 16:02:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:32.101 16:02:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:32.101 16:02:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:32.101 "name": "raid_bdev1", 00:14:32.101 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:32.101 "strip_size_kb": 0, 00:14:32.101 "state": "online", 00:14:32.101 "raid_level": "raid1", 00:14:32.101 "superblock": true, 00:14:32.101 "num_base_bdevs": 4, 00:14:32.101 "num_base_bdevs_discovered": 3, 00:14:32.101 "num_base_bdevs_operational": 3, 00:14:32.101 "process": { 00:14:32.101 "type": "rebuild", 00:14:32.101 "target": "spare", 00:14:32.101 "progress": { 00:14:32.101 "blocks": 34816, 00:14:32.101 "percent": 54 00:14:32.101 } 00:14:32.101 }, 00:14:32.101 "base_bdevs_list": [ 00:14:32.101 { 00:14:32.101 "name": "spare", 00:14:32.101 "uuid": "9b4e9028-25d6-50e1-8f5c-fcb2012fb089", 00:14:32.101 "is_configured": true, 00:14:32.101 "data_offset": 2048, 00:14:32.101 "data_size": 63488 00:14:32.101 }, 00:14:32.102 { 00:14:32.102 "name": null, 00:14:32.102 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:32.102 "is_configured": false, 00:14:32.102 "data_offset": 0, 00:14:32.102 "data_size": 63488 00:14:32.102 }, 00:14:32.102 { 00:14:32.102 "name": "BaseBdev3", 00:14:32.102 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:32.102 "is_configured": true, 00:14:32.102 "data_offset": 2048, 00:14:32.102 "data_size": 63488 00:14:32.102 }, 00:14:32.102 { 00:14:32.102 "name": "BaseBdev4", 00:14:32.102 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:32.102 "is_configured": true, 00:14:32.102 "data_offset": 2048, 00:14:32.102 "data_size": 63488 00:14:32.102 } 00:14:32.102 ] 00:14:32.102 }' 00:14:32.102 16:02:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:32.102 16:02:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:32.102 16:02:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:32.102 16:02:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:32.102 16:02:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:32.718 103.20 IOPS, 309.60 MiB/s [2024-11-20T16:02:30.969Z] [2024-11-20 16:02:30.687719] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:14:33.001 16:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:33.001 16:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:33.001 16:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:33.001 16:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:33.001 16:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:33.001 16:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:33.001 16:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.001 16:02:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.001 16:02:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:33.001 16:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:33.001 16:02:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.001 16:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:33.001 "name": "raid_bdev1", 00:14:33.001 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:33.001 "strip_size_kb": 0, 00:14:33.001 "state": "online", 00:14:33.001 "raid_level": "raid1", 00:14:33.001 "superblock": true, 00:14:33.001 "num_base_bdevs": 4, 00:14:33.001 "num_base_bdevs_discovered": 3, 00:14:33.001 "num_base_bdevs_operational": 3, 00:14:33.001 "process": { 00:14:33.001 "type": "rebuild", 00:14:33.001 "target": "spare", 00:14:33.001 "progress": { 00:14:33.001 "blocks": 55296, 00:14:33.001 "percent": 87 00:14:33.001 } 00:14:33.001 }, 00:14:33.001 "base_bdevs_list": [ 00:14:33.001 { 00:14:33.001 "name": "spare", 00:14:33.001 "uuid": "9b4e9028-25d6-50e1-8f5c-fcb2012fb089", 00:14:33.001 "is_configured": true, 00:14:33.001 "data_offset": 2048, 00:14:33.001 "data_size": 63488 00:14:33.001 }, 00:14:33.001 { 00:14:33.001 "name": null, 00:14:33.001 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.001 "is_configured": false, 00:14:33.001 "data_offset": 0, 00:14:33.001 "data_size": 63488 00:14:33.001 }, 00:14:33.001 { 00:14:33.001 "name": "BaseBdev3", 00:14:33.001 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:33.001 "is_configured": true, 00:14:33.001 "data_offset": 2048, 00:14:33.001 "data_size": 63488 00:14:33.001 }, 00:14:33.001 { 00:14:33.001 "name": "BaseBdev4", 00:14:33.001 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:33.001 "is_configured": true, 00:14:33.001 "data_offset": 2048, 00:14:33.001 "data_size": 63488 00:14:33.001 } 00:14:33.001 ] 00:14:33.001 }' 00:14:33.001 16:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:33.001 16:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:33.001 16:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:33.001 [2024-11-20 16:02:31.240693] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:14:33.261 16:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:33.261 16:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:33.261 94.33 IOPS, 283.00 MiB/s [2024-11-20T16:02:31.512Z] [2024-11-20 16:02:31.444102] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:14:33.261 [2024-11-20 16:02:31.444541] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:14:33.833 [2024-11-20 16:02:31.788560] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:33.833 [2024-11-20 16:02:31.888567] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:33.833 [2024-11-20 16:02:31.890836] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:34.095 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:34.095 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:34.095 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:34.095 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:34.095 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:34.095 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:34.095 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:34.095 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.095 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:34.095 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:34.095 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.095 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:34.095 "name": "raid_bdev1", 00:14:34.095 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:34.095 "strip_size_kb": 0, 00:14:34.095 "state": "online", 00:14:34.095 "raid_level": "raid1", 00:14:34.095 "superblock": true, 00:14:34.095 "num_base_bdevs": 4, 00:14:34.095 "num_base_bdevs_discovered": 3, 00:14:34.095 "num_base_bdevs_operational": 3, 00:14:34.095 "base_bdevs_list": [ 00:14:34.095 { 00:14:34.095 "name": "spare", 00:14:34.095 "uuid": "9b4e9028-25d6-50e1-8f5c-fcb2012fb089", 00:14:34.095 "is_configured": true, 00:14:34.095 "data_offset": 2048, 00:14:34.095 "data_size": 63488 00:14:34.095 }, 00:14:34.095 { 00:14:34.095 "name": null, 00:14:34.095 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:34.095 "is_configured": false, 00:14:34.095 "data_offset": 0, 00:14:34.095 "data_size": 63488 00:14:34.095 }, 00:14:34.095 { 00:14:34.095 "name": "BaseBdev3", 00:14:34.095 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:34.095 "is_configured": true, 00:14:34.095 "data_offset": 2048, 00:14:34.095 "data_size": 63488 00:14:34.095 }, 00:14:34.095 { 00:14:34.095 "name": "BaseBdev4", 00:14:34.095 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:34.095 "is_configured": true, 00:14:34.095 "data_offset": 2048, 00:14:34.095 "data_size": 63488 00:14:34.095 } 00:14:34.095 ] 00:14:34.095 }' 00:14:34.095 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:34.356 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:34.356 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:34.356 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:34.356 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:14:34.356 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:34.356 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:34.356 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:34.356 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:34.356 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:34.356 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:34.356 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:34.356 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.356 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:34.356 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.356 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:34.356 "name": "raid_bdev1", 00:14:34.356 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:34.356 "strip_size_kb": 0, 00:14:34.356 "state": "online", 00:14:34.356 "raid_level": "raid1", 00:14:34.356 "superblock": true, 00:14:34.356 "num_base_bdevs": 4, 00:14:34.356 "num_base_bdevs_discovered": 3, 00:14:34.356 "num_base_bdevs_operational": 3, 00:14:34.356 "base_bdevs_list": [ 00:14:34.356 { 00:14:34.356 "name": "spare", 00:14:34.356 "uuid": "9b4e9028-25d6-50e1-8f5c-fcb2012fb089", 00:14:34.356 "is_configured": true, 00:14:34.356 "data_offset": 2048, 00:14:34.356 "data_size": 63488 00:14:34.356 }, 00:14:34.356 { 00:14:34.356 "name": null, 00:14:34.356 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:34.356 "is_configured": false, 00:14:34.356 "data_offset": 0, 00:14:34.356 "data_size": 63488 00:14:34.356 }, 00:14:34.356 { 00:14:34.356 "name": "BaseBdev3", 00:14:34.356 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:34.356 "is_configured": true, 00:14:34.356 "data_offset": 2048, 00:14:34.356 "data_size": 63488 00:14:34.356 }, 00:14:34.356 { 00:14:34.356 "name": "BaseBdev4", 00:14:34.356 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:34.356 "is_configured": true, 00:14:34.357 "data_offset": 2048, 00:14:34.357 "data_size": 63488 00:14:34.357 } 00:14:34.357 ] 00:14:34.357 }' 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:34.357 85.43 IOPS, 256.29 MiB/s [2024-11-20T16:02:32.608Z] 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:34.357 "name": "raid_bdev1", 00:14:34.357 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:34.357 "strip_size_kb": 0, 00:14:34.357 "state": "online", 00:14:34.357 "raid_level": "raid1", 00:14:34.357 "superblock": true, 00:14:34.357 "num_base_bdevs": 4, 00:14:34.357 "num_base_bdevs_discovered": 3, 00:14:34.357 "num_base_bdevs_operational": 3, 00:14:34.357 "base_bdevs_list": [ 00:14:34.357 { 00:14:34.357 "name": "spare", 00:14:34.357 "uuid": "9b4e9028-25d6-50e1-8f5c-fcb2012fb089", 00:14:34.357 "is_configured": true, 00:14:34.357 "data_offset": 2048, 00:14:34.357 "data_size": 63488 00:14:34.357 }, 00:14:34.357 { 00:14:34.357 "name": null, 00:14:34.357 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:34.357 "is_configured": false, 00:14:34.357 "data_offset": 0, 00:14:34.357 "data_size": 63488 00:14:34.357 }, 00:14:34.357 { 00:14:34.357 "name": "BaseBdev3", 00:14:34.357 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:34.357 "is_configured": true, 00:14:34.357 "data_offset": 2048, 00:14:34.357 "data_size": 63488 00:14:34.357 }, 00:14:34.357 { 00:14:34.357 "name": "BaseBdev4", 00:14:34.357 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:34.357 "is_configured": true, 00:14:34.357 "data_offset": 2048, 00:14:34.357 "data_size": 63488 00:14:34.357 } 00:14:34.357 ] 00:14:34.357 }' 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:34.357 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:34.617 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:34.617 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.617 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:34.617 [2024-11-20 16:02:32.845364] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:34.617 [2024-11-20 16:02:32.845394] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:34.876 00:14:34.877 Latency(us) 00:14:34.877 [2024-11-20T16:02:33.128Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:34.877 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:14:34.877 raid_bdev1 : 7.46 82.02 246.06 0.00 0.00 16833.54 308.78 111310.38 00:14:34.877 [2024-11-20T16:02:33.128Z] =================================================================================================================== 00:14:34.877 [2024-11-20T16:02:33.128Z] Total : 82.02 246.06 0.00 0.00 16833.54 308.78 111310.38 00:14:34.877 { 00:14:34.877 "results": [ 00:14:34.877 { 00:14:34.877 "job": "raid_bdev1", 00:14:34.877 "core_mask": "0x1", 00:14:34.877 "workload": "randrw", 00:14:34.877 "percentage": 50, 00:14:34.877 "status": "finished", 00:14:34.877 "queue_depth": 2, 00:14:34.877 "io_size": 3145728, 00:14:34.877 "runtime": 7.461499, 00:14:34.877 "iops": 82.02105233814278, 00:14:34.877 "mibps": 246.06315701442833, 00:14:34.877 "io_failed": 0, 00:14:34.877 "io_timeout": 0, 00:14:34.877 "avg_latency_us": 16833.544494720965, 00:14:34.877 "min_latency_us": 308.7753846153846, 00:14:34.877 "max_latency_us": 111310.37538461538 00:14:34.877 } 00:14:34.877 ], 00:14:34.877 "core_count": 1 00:14:34.877 } 00:14:34.877 [2024-11-20 16:02:32.892954] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:34.877 [2024-11-20 16:02:32.893020] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:34.877 [2024-11-20 16:02:32.893125] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:34.877 [2024-11-20 16:02:32.893135] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:34.877 16:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:14:35.136 /dev/nbd0 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:35.136 1+0 records in 00:14:35.136 1+0 records out 00:14:35.136 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000271926 s, 15.1 MB/s 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@728 -- # continue 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:14:35.136 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:35.137 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:14:35.137 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:35.137 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:14:35.137 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:35.137 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:14:35.137 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:35.137 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:35.137 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:14:35.137 /dev/nbd1 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:35.397 1+0 records in 00:14:35.397 1+0 records out 00:14:35.397 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00025288 s, 16.2 MB/s 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:35.397 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:35.660 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:14:35.922 /dev/nbd1 00:14:35.922 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:35.922 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:35.922 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:14:35.922 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:14:35.922 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:35.922 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:35.922 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:14:35.922 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:14:35.922 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:35.922 16:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:35.922 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:35.922 1+0 records in 00:14:35.922 1+0 records out 00:14:35.922 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000346861 s, 11.8 MB/s 00:14:35.922 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:35.922 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:14:35.922 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:35.923 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:35.923 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:14:35.923 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:35.923 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:35.923 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:14:35.923 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:14:35.923 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:35.923 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:14:35.923 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:35.923 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:14:35.923 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:35.923 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:36.184 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:36.184 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:36.184 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:36.184 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:36.184 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:36.184 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:36.184 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:14:36.184 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:14:36.184 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:36.184 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:36.184 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:36.184 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:36.184 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:14:36.184 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:36.184 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:36.445 [2024-11-20 16:02:34.548615] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:36.445 [2024-11-20 16:02:34.548687] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:36.445 [2024-11-20 16:02:34.548721] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:14:36.445 [2024-11-20 16:02:34.548732] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:36.445 [2024-11-20 16:02:34.551020] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:36.445 [2024-11-20 16:02:34.551166] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:36.445 [2024-11-20 16:02:34.551271] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:36.445 [2024-11-20 16:02:34.551320] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:36.445 [2024-11-20 16:02:34.551460] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:36.445 [2024-11-20 16:02:34.551557] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:36.445 spare 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:36.445 [2024-11-20 16:02:34.651658] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:14:36.445 [2024-11-20 16:02:34.651705] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:14:36.445 [2024-11-20 16:02:34.652059] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000037160 00:14:36.445 [2024-11-20 16:02:34.652276] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:14:36.445 [2024-11-20 16:02:34.652295] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:14:36.445 [2024-11-20 16:02:34.652493] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.445 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.446 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:36.446 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:36.446 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.446 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:36.446 "name": "raid_bdev1", 00:14:36.446 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:36.446 "strip_size_kb": 0, 00:14:36.446 "state": "online", 00:14:36.446 "raid_level": "raid1", 00:14:36.446 "superblock": true, 00:14:36.446 "num_base_bdevs": 4, 00:14:36.446 "num_base_bdevs_discovered": 3, 00:14:36.446 "num_base_bdevs_operational": 3, 00:14:36.446 "base_bdevs_list": [ 00:14:36.446 { 00:14:36.446 "name": "spare", 00:14:36.446 "uuid": "9b4e9028-25d6-50e1-8f5c-fcb2012fb089", 00:14:36.446 "is_configured": true, 00:14:36.446 "data_offset": 2048, 00:14:36.446 "data_size": 63488 00:14:36.446 }, 00:14:36.446 { 00:14:36.446 "name": null, 00:14:36.446 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:36.446 "is_configured": false, 00:14:36.446 "data_offset": 2048, 00:14:36.446 "data_size": 63488 00:14:36.446 }, 00:14:36.446 { 00:14:36.446 "name": "BaseBdev3", 00:14:36.446 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:36.446 "is_configured": true, 00:14:36.446 "data_offset": 2048, 00:14:36.446 "data_size": 63488 00:14:36.446 }, 00:14:36.446 { 00:14:36.446 "name": "BaseBdev4", 00:14:36.446 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:36.446 "is_configured": true, 00:14:36.446 "data_offset": 2048, 00:14:36.446 "data_size": 63488 00:14:36.446 } 00:14:36.446 ] 00:14:36.446 }' 00:14:36.446 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:36.446 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:37.019 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:37.019 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:37.019 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:37.019 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:37.019 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:37.019 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.019 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:37.019 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.019 16:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:37.019 "name": "raid_bdev1", 00:14:37.019 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:37.019 "strip_size_kb": 0, 00:14:37.019 "state": "online", 00:14:37.019 "raid_level": "raid1", 00:14:37.019 "superblock": true, 00:14:37.019 "num_base_bdevs": 4, 00:14:37.019 "num_base_bdevs_discovered": 3, 00:14:37.019 "num_base_bdevs_operational": 3, 00:14:37.019 "base_bdevs_list": [ 00:14:37.019 { 00:14:37.019 "name": "spare", 00:14:37.019 "uuid": "9b4e9028-25d6-50e1-8f5c-fcb2012fb089", 00:14:37.019 "is_configured": true, 00:14:37.019 "data_offset": 2048, 00:14:37.019 "data_size": 63488 00:14:37.019 }, 00:14:37.019 { 00:14:37.019 "name": null, 00:14:37.019 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:37.019 "is_configured": false, 00:14:37.019 "data_offset": 2048, 00:14:37.019 "data_size": 63488 00:14:37.019 }, 00:14:37.019 { 00:14:37.019 "name": "BaseBdev3", 00:14:37.019 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:37.019 "is_configured": true, 00:14:37.019 "data_offset": 2048, 00:14:37.019 "data_size": 63488 00:14:37.019 }, 00:14:37.019 { 00:14:37.019 "name": "BaseBdev4", 00:14:37.019 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:37.019 "is_configured": true, 00:14:37.019 "data_offset": 2048, 00:14:37.019 "data_size": 63488 00:14:37.019 } 00:14:37.019 ] 00:14:37.019 }' 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:37.019 [2024-11-20 16:02:35.132914] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:37.019 "name": "raid_bdev1", 00:14:37.019 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:37.019 "strip_size_kb": 0, 00:14:37.019 "state": "online", 00:14:37.019 "raid_level": "raid1", 00:14:37.019 "superblock": true, 00:14:37.019 "num_base_bdevs": 4, 00:14:37.019 "num_base_bdevs_discovered": 2, 00:14:37.019 "num_base_bdevs_operational": 2, 00:14:37.019 "base_bdevs_list": [ 00:14:37.019 { 00:14:37.019 "name": null, 00:14:37.019 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:37.019 "is_configured": false, 00:14:37.019 "data_offset": 0, 00:14:37.019 "data_size": 63488 00:14:37.019 }, 00:14:37.019 { 00:14:37.019 "name": null, 00:14:37.019 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:37.019 "is_configured": false, 00:14:37.019 "data_offset": 2048, 00:14:37.019 "data_size": 63488 00:14:37.019 }, 00:14:37.019 { 00:14:37.019 "name": "BaseBdev3", 00:14:37.019 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:37.019 "is_configured": true, 00:14:37.019 "data_offset": 2048, 00:14:37.019 "data_size": 63488 00:14:37.019 }, 00:14:37.019 { 00:14:37.019 "name": "BaseBdev4", 00:14:37.019 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:37.019 "is_configured": true, 00:14:37.019 "data_offset": 2048, 00:14:37.019 "data_size": 63488 00:14:37.019 } 00:14:37.019 ] 00:14:37.019 }' 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:37.019 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:37.279 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:37.279 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.279 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:37.279 [2024-11-20 16:02:35.461078] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:37.280 [2024-11-20 16:02:35.461376] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:14:37.280 [2024-11-20 16:02:35.461396] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:37.280 [2024-11-20 16:02:35.461434] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:37.280 [2024-11-20 16:02:35.471280] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000037230 00:14:37.280 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.280 16:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:14:37.280 [2024-11-20 16:02:35.473259] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:38.664 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:38.665 "name": "raid_bdev1", 00:14:38.665 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:38.665 "strip_size_kb": 0, 00:14:38.665 "state": "online", 00:14:38.665 "raid_level": "raid1", 00:14:38.665 "superblock": true, 00:14:38.665 "num_base_bdevs": 4, 00:14:38.665 "num_base_bdevs_discovered": 3, 00:14:38.665 "num_base_bdevs_operational": 3, 00:14:38.665 "process": { 00:14:38.665 "type": "rebuild", 00:14:38.665 "target": "spare", 00:14:38.665 "progress": { 00:14:38.665 "blocks": 20480, 00:14:38.665 "percent": 32 00:14:38.665 } 00:14:38.665 }, 00:14:38.665 "base_bdevs_list": [ 00:14:38.665 { 00:14:38.665 "name": "spare", 00:14:38.665 "uuid": "9b4e9028-25d6-50e1-8f5c-fcb2012fb089", 00:14:38.665 "is_configured": true, 00:14:38.665 "data_offset": 2048, 00:14:38.665 "data_size": 63488 00:14:38.665 }, 00:14:38.665 { 00:14:38.665 "name": null, 00:14:38.665 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:38.665 "is_configured": false, 00:14:38.665 "data_offset": 2048, 00:14:38.665 "data_size": 63488 00:14:38.665 }, 00:14:38.665 { 00:14:38.665 "name": "BaseBdev3", 00:14:38.665 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:38.665 "is_configured": true, 00:14:38.665 "data_offset": 2048, 00:14:38.665 "data_size": 63488 00:14:38.665 }, 00:14:38.665 { 00:14:38.665 "name": "BaseBdev4", 00:14:38.665 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:38.665 "is_configured": true, 00:14:38.665 "data_offset": 2048, 00:14:38.665 "data_size": 63488 00:14:38.665 } 00:14:38.665 ] 00:14:38.665 }' 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:38.665 [2024-11-20 16:02:36.587135] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:38.665 [2024-11-20 16:02:36.679565] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:38.665 [2024-11-20 16:02:36.679648] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:38.665 [2024-11-20 16:02:36.679686] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:38.665 [2024-11-20 16:02:36.679695] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:38.665 "name": "raid_bdev1", 00:14:38.665 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:38.665 "strip_size_kb": 0, 00:14:38.665 "state": "online", 00:14:38.665 "raid_level": "raid1", 00:14:38.665 "superblock": true, 00:14:38.665 "num_base_bdevs": 4, 00:14:38.665 "num_base_bdevs_discovered": 2, 00:14:38.665 "num_base_bdevs_operational": 2, 00:14:38.665 "base_bdevs_list": [ 00:14:38.665 { 00:14:38.665 "name": null, 00:14:38.665 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:38.665 "is_configured": false, 00:14:38.665 "data_offset": 0, 00:14:38.665 "data_size": 63488 00:14:38.665 }, 00:14:38.665 { 00:14:38.665 "name": null, 00:14:38.665 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:38.665 "is_configured": false, 00:14:38.665 "data_offset": 2048, 00:14:38.665 "data_size": 63488 00:14:38.665 }, 00:14:38.665 { 00:14:38.665 "name": "BaseBdev3", 00:14:38.665 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:38.665 "is_configured": true, 00:14:38.665 "data_offset": 2048, 00:14:38.665 "data_size": 63488 00:14:38.665 }, 00:14:38.665 { 00:14:38.665 "name": "BaseBdev4", 00:14:38.665 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:38.665 "is_configured": true, 00:14:38.665 "data_offset": 2048, 00:14:38.665 "data_size": 63488 00:14:38.665 } 00:14:38.665 ] 00:14:38.665 }' 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:38.665 16:02:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:38.925 16:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:38.925 16:02:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.925 16:02:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:38.925 [2024-11-20 16:02:37.064364] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:38.925 [2024-11-20 16:02:37.064424] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:38.925 [2024-11-20 16:02:37.064453] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:14:38.925 [2024-11-20 16:02:37.064462] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:38.925 [2024-11-20 16:02:37.065027] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:38.925 [2024-11-20 16:02:37.065060] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:38.925 [2024-11-20 16:02:37.065174] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:38.925 [2024-11-20 16:02:37.065188] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:14:38.925 [2024-11-20 16:02:37.065199] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:38.925 [2024-11-20 16:02:37.065226] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:38.925 [2024-11-20 16:02:37.074771] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000037300 00:14:38.925 spare 00:14:38.925 16:02:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.925 16:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:14:38.925 [2024-11-20 16:02:37.076804] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:39.863 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:39.863 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:39.863 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:39.863 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:39.863 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:39.863 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.863 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:39.863 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.863 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:39.863 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:40.124 "name": "raid_bdev1", 00:14:40.124 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:40.124 "strip_size_kb": 0, 00:14:40.124 "state": "online", 00:14:40.124 "raid_level": "raid1", 00:14:40.124 "superblock": true, 00:14:40.124 "num_base_bdevs": 4, 00:14:40.124 "num_base_bdevs_discovered": 3, 00:14:40.124 "num_base_bdevs_operational": 3, 00:14:40.124 "process": { 00:14:40.124 "type": "rebuild", 00:14:40.124 "target": "spare", 00:14:40.124 "progress": { 00:14:40.124 "blocks": 20480, 00:14:40.124 "percent": 32 00:14:40.124 } 00:14:40.124 }, 00:14:40.124 "base_bdevs_list": [ 00:14:40.124 { 00:14:40.124 "name": "spare", 00:14:40.124 "uuid": "9b4e9028-25d6-50e1-8f5c-fcb2012fb089", 00:14:40.124 "is_configured": true, 00:14:40.124 "data_offset": 2048, 00:14:40.124 "data_size": 63488 00:14:40.124 }, 00:14:40.124 { 00:14:40.124 "name": null, 00:14:40.124 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:40.124 "is_configured": false, 00:14:40.124 "data_offset": 2048, 00:14:40.124 "data_size": 63488 00:14:40.124 }, 00:14:40.124 { 00:14:40.124 "name": "BaseBdev3", 00:14:40.124 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:40.124 "is_configured": true, 00:14:40.124 "data_offset": 2048, 00:14:40.124 "data_size": 63488 00:14:40.124 }, 00:14:40.124 { 00:14:40.124 "name": "BaseBdev4", 00:14:40.124 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:40.124 "is_configured": true, 00:14:40.124 "data_offset": 2048, 00:14:40.124 "data_size": 63488 00:14:40.124 } 00:14:40.124 ] 00:14:40.124 }' 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:40.124 [2024-11-20 16:02:38.186682] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:40.124 [2024-11-20 16:02:38.283071] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:40.124 [2024-11-20 16:02:38.283148] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:40.124 [2024-11-20 16:02:38.283165] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:40.124 [2024-11-20 16:02:38.283177] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.124 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:40.124 "name": "raid_bdev1", 00:14:40.124 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:40.124 "strip_size_kb": 0, 00:14:40.124 "state": "online", 00:14:40.124 "raid_level": "raid1", 00:14:40.124 "superblock": true, 00:14:40.124 "num_base_bdevs": 4, 00:14:40.124 "num_base_bdevs_discovered": 2, 00:14:40.124 "num_base_bdevs_operational": 2, 00:14:40.124 "base_bdevs_list": [ 00:14:40.124 { 00:14:40.124 "name": null, 00:14:40.125 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:40.125 "is_configured": false, 00:14:40.125 "data_offset": 0, 00:14:40.125 "data_size": 63488 00:14:40.125 }, 00:14:40.125 { 00:14:40.125 "name": null, 00:14:40.125 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:40.125 "is_configured": false, 00:14:40.125 "data_offset": 2048, 00:14:40.125 "data_size": 63488 00:14:40.125 }, 00:14:40.125 { 00:14:40.125 "name": "BaseBdev3", 00:14:40.125 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:40.125 "is_configured": true, 00:14:40.125 "data_offset": 2048, 00:14:40.125 "data_size": 63488 00:14:40.125 }, 00:14:40.125 { 00:14:40.125 "name": "BaseBdev4", 00:14:40.125 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:40.125 "is_configured": true, 00:14:40.125 "data_offset": 2048, 00:14:40.125 "data_size": 63488 00:14:40.125 } 00:14:40.125 ] 00:14:40.125 }' 00:14:40.125 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:40.125 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:40.384 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:40.384 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:40.384 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:40.384 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:40.384 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:40.384 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.384 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:40.384 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.384 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:40.643 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.643 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:40.643 "name": "raid_bdev1", 00:14:40.643 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:40.643 "strip_size_kb": 0, 00:14:40.643 "state": "online", 00:14:40.643 "raid_level": "raid1", 00:14:40.643 "superblock": true, 00:14:40.643 "num_base_bdevs": 4, 00:14:40.643 "num_base_bdevs_discovered": 2, 00:14:40.643 "num_base_bdevs_operational": 2, 00:14:40.643 "base_bdevs_list": [ 00:14:40.643 { 00:14:40.643 "name": null, 00:14:40.643 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:40.643 "is_configured": false, 00:14:40.643 "data_offset": 0, 00:14:40.643 "data_size": 63488 00:14:40.643 }, 00:14:40.643 { 00:14:40.643 "name": null, 00:14:40.643 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:40.643 "is_configured": false, 00:14:40.643 "data_offset": 2048, 00:14:40.643 "data_size": 63488 00:14:40.643 }, 00:14:40.643 { 00:14:40.643 "name": "BaseBdev3", 00:14:40.643 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:40.643 "is_configured": true, 00:14:40.643 "data_offset": 2048, 00:14:40.643 "data_size": 63488 00:14:40.643 }, 00:14:40.643 { 00:14:40.643 "name": "BaseBdev4", 00:14:40.643 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:40.643 "is_configured": true, 00:14:40.643 "data_offset": 2048, 00:14:40.643 "data_size": 63488 00:14:40.643 } 00:14:40.643 ] 00:14:40.643 }' 00:14:40.643 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:40.643 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:40.643 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:40.643 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:40.643 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:14:40.643 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.643 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:40.643 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.643 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:40.643 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.643 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:40.643 [2024-11-20 16:02:38.727503] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:40.643 [2024-11-20 16:02:38.727561] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:40.643 [2024-11-20 16:02:38.727580] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000cc80 00:14:40.643 [2024-11-20 16:02:38.727590] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:40.643 [2024-11-20 16:02:38.728024] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:40.643 [2024-11-20 16:02:38.728046] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:40.643 [2024-11-20 16:02:38.728120] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:14:40.643 [2024-11-20 16:02:38.728138] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:14:40.643 [2024-11-20 16:02:38.728146] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:40.643 [2024-11-20 16:02:38.728158] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:14:40.643 BaseBdev1 00:14:40.643 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.643 16:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:41.583 "name": "raid_bdev1", 00:14:41.583 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:41.583 "strip_size_kb": 0, 00:14:41.583 "state": "online", 00:14:41.583 "raid_level": "raid1", 00:14:41.583 "superblock": true, 00:14:41.583 "num_base_bdevs": 4, 00:14:41.583 "num_base_bdevs_discovered": 2, 00:14:41.583 "num_base_bdevs_operational": 2, 00:14:41.583 "base_bdevs_list": [ 00:14:41.583 { 00:14:41.583 "name": null, 00:14:41.583 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:41.583 "is_configured": false, 00:14:41.583 "data_offset": 0, 00:14:41.583 "data_size": 63488 00:14:41.583 }, 00:14:41.583 { 00:14:41.583 "name": null, 00:14:41.583 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:41.583 "is_configured": false, 00:14:41.583 "data_offset": 2048, 00:14:41.583 "data_size": 63488 00:14:41.583 }, 00:14:41.583 { 00:14:41.583 "name": "BaseBdev3", 00:14:41.583 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:41.583 "is_configured": true, 00:14:41.583 "data_offset": 2048, 00:14:41.583 "data_size": 63488 00:14:41.583 }, 00:14:41.583 { 00:14:41.583 "name": "BaseBdev4", 00:14:41.583 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:41.583 "is_configured": true, 00:14:41.583 "data_offset": 2048, 00:14:41.583 "data_size": 63488 00:14:41.583 } 00:14:41.583 ] 00:14:41.583 }' 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:41.583 16:02:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:41.902 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:41.902 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:41.902 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:41.902 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:41.902 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:41.902 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:41.902 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.902 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.902 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:41.902 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.902 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:41.902 "name": "raid_bdev1", 00:14:41.902 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:41.902 "strip_size_kb": 0, 00:14:41.902 "state": "online", 00:14:41.902 "raid_level": "raid1", 00:14:41.902 "superblock": true, 00:14:41.902 "num_base_bdevs": 4, 00:14:41.902 "num_base_bdevs_discovered": 2, 00:14:41.902 "num_base_bdevs_operational": 2, 00:14:41.902 "base_bdevs_list": [ 00:14:41.902 { 00:14:41.902 "name": null, 00:14:41.902 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:41.902 "is_configured": false, 00:14:41.902 "data_offset": 0, 00:14:41.902 "data_size": 63488 00:14:41.902 }, 00:14:41.902 { 00:14:41.902 "name": null, 00:14:41.902 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:41.902 "is_configured": false, 00:14:41.902 "data_offset": 2048, 00:14:41.902 "data_size": 63488 00:14:41.902 }, 00:14:41.902 { 00:14:41.902 "name": "BaseBdev3", 00:14:41.902 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:41.902 "is_configured": true, 00:14:41.902 "data_offset": 2048, 00:14:41.902 "data_size": 63488 00:14:41.902 }, 00:14:41.902 { 00:14:41.902 "name": "BaseBdev4", 00:14:41.902 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:41.902 "is_configured": true, 00:14:41.902 "data_offset": 2048, 00:14:41.902 "data_size": 63488 00:14:41.902 } 00:14:41.902 ] 00:14:41.902 }' 00:14:41.902 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:42.162 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:42.162 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:42.162 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:42.162 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:42.162 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # local es=0 00:14:42.162 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:42.162 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:14:42.162 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:42.162 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:14:42.162 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:42.162 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:42.162 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.162 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:42.162 [2024-11-20 16:02:40.172105] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:42.162 [2024-11-20 16:02:40.172265] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:14:42.162 [2024-11-20 16:02:40.172280] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:42.162 request: 00:14:42.162 { 00:14:42.162 "base_bdev": "BaseBdev1", 00:14:42.162 "raid_bdev": "raid_bdev1", 00:14:42.162 "method": "bdev_raid_add_base_bdev", 00:14:42.162 "req_id": 1 00:14:42.162 } 00:14:42.162 Got JSON-RPC error response 00:14:42.162 response: 00:14:42.162 { 00:14:42.162 "code": -22, 00:14:42.162 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:14:42.162 } 00:14:42.162 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:14:42.163 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # es=1 00:14:42.163 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:42.163 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:42.163 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:42.163 16:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:43.103 "name": "raid_bdev1", 00:14:43.103 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:43.103 "strip_size_kb": 0, 00:14:43.103 "state": "online", 00:14:43.103 "raid_level": "raid1", 00:14:43.103 "superblock": true, 00:14:43.103 "num_base_bdevs": 4, 00:14:43.103 "num_base_bdevs_discovered": 2, 00:14:43.103 "num_base_bdevs_operational": 2, 00:14:43.103 "base_bdevs_list": [ 00:14:43.103 { 00:14:43.103 "name": null, 00:14:43.103 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:43.103 "is_configured": false, 00:14:43.103 "data_offset": 0, 00:14:43.103 "data_size": 63488 00:14:43.103 }, 00:14:43.103 { 00:14:43.103 "name": null, 00:14:43.103 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:43.103 "is_configured": false, 00:14:43.103 "data_offset": 2048, 00:14:43.103 "data_size": 63488 00:14:43.103 }, 00:14:43.103 { 00:14:43.103 "name": "BaseBdev3", 00:14:43.103 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:43.103 "is_configured": true, 00:14:43.103 "data_offset": 2048, 00:14:43.103 "data_size": 63488 00:14:43.103 }, 00:14:43.103 { 00:14:43.103 "name": "BaseBdev4", 00:14:43.103 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:43.103 "is_configured": true, 00:14:43.103 "data_offset": 2048, 00:14:43.103 "data_size": 63488 00:14:43.103 } 00:14:43.103 ] 00:14:43.103 }' 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:43.103 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:43.363 "name": "raid_bdev1", 00:14:43.363 "uuid": "ca597bef-3e46-4661-a09a-faf8784ebf0e", 00:14:43.363 "strip_size_kb": 0, 00:14:43.363 "state": "online", 00:14:43.363 "raid_level": "raid1", 00:14:43.363 "superblock": true, 00:14:43.363 "num_base_bdevs": 4, 00:14:43.363 "num_base_bdevs_discovered": 2, 00:14:43.363 "num_base_bdevs_operational": 2, 00:14:43.363 "base_bdevs_list": [ 00:14:43.363 { 00:14:43.363 "name": null, 00:14:43.363 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:43.363 "is_configured": false, 00:14:43.363 "data_offset": 0, 00:14:43.363 "data_size": 63488 00:14:43.363 }, 00:14:43.363 { 00:14:43.363 "name": null, 00:14:43.363 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:43.363 "is_configured": false, 00:14:43.363 "data_offset": 2048, 00:14:43.363 "data_size": 63488 00:14:43.363 }, 00:14:43.363 { 00:14:43.363 "name": "BaseBdev3", 00:14:43.363 "uuid": "91858571-6d69-57c0-8ce6-f4c4d38d6829", 00:14:43.363 "is_configured": true, 00:14:43.363 "data_offset": 2048, 00:14:43.363 "data_size": 63488 00:14:43.363 }, 00:14:43.363 { 00:14:43.363 "name": "BaseBdev4", 00:14:43.363 "uuid": "95bd6472-9999-54a6-ac8d-6ae19ea077ae", 00:14:43.363 "is_configured": true, 00:14:43.363 "data_offset": 2048, 00:14:43.363 "data_size": 63488 00:14:43.363 } 00:14:43.363 ] 00:14:43.363 }' 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 77026 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # '[' -z 77026 ']' 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # kill -0 77026 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # uname 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:43.363 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 77026 00:14:43.622 killing process with pid 77026 00:14:43.622 Received shutdown signal, test time was about 16.200267 seconds 00:14:43.622 00:14:43.622 Latency(us) 00:14:43.622 [2024-11-20T16:02:41.873Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:43.622 [2024-11-20T16:02:41.873Z] =================================================================================================================== 00:14:43.622 [2024-11-20T16:02:41.873Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:14:43.622 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:43.622 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:43.622 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 77026' 00:14:43.622 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@973 -- # kill 77026 00:14:43.622 16:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@978 -- # wait 77026 00:14:43.623 [2024-11-20 16:02:41.616111] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:43.623 [2024-11-20 16:02:41.616235] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:43.623 [2024-11-20 16:02:41.616312] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:43.623 [2024-11-20 16:02:41.616322] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:14:43.883 [2024-11-20 16:02:41.963066] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:44.820 16:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:14:44.820 00:14:44.820 real 0m18.921s 00:14:44.820 user 0m24.110s 00:14:44.820 sys 0m1.719s 00:14:44.820 ************************************ 00:14:44.820 END TEST raid_rebuild_test_sb_io 00:14:44.820 ************************************ 00:14:44.820 16:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:44.820 16:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:44.821 16:02:42 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:14:44.821 16:02:42 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 3 false 00:14:44.821 16:02:42 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:14:44.821 16:02:42 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:44.821 16:02:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:44.821 ************************************ 00:14:44.821 START TEST raid5f_state_function_test 00:14:44.821 ************************************ 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 3 false 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:44.821 Process raid pid: 77728 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=77728 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 77728' 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 77728 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 77728 ']' 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:44.821 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.821 16:02:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:44.821 [2024-11-20 16:02:42.855069] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:14:44.821 [2024-11-20 16:02:42.855204] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:44.821 [2024-11-20 16:02:43.014392] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:45.081 [2024-11-20 16:02:43.117689] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:45.081 [2024-11-20 16:02:43.256261] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:45.081 [2024-11-20 16:02:43.256298] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.652 [2024-11-20 16:02:43.705550] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:45.652 [2024-11-20 16:02:43.705599] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:45.652 [2024-11-20 16:02:43.705609] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:45.652 [2024-11-20 16:02:43.705619] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:45.652 [2024-11-20 16:02:43.705625] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:45.652 [2024-11-20 16:02:43.705634] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.652 16:02:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:45.652 "name": "Existed_Raid", 00:14:45.652 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:45.652 "strip_size_kb": 64, 00:14:45.652 "state": "configuring", 00:14:45.652 "raid_level": "raid5f", 00:14:45.652 "superblock": false, 00:14:45.652 "num_base_bdevs": 3, 00:14:45.652 "num_base_bdevs_discovered": 0, 00:14:45.652 "num_base_bdevs_operational": 3, 00:14:45.652 "base_bdevs_list": [ 00:14:45.652 { 00:14:45.652 "name": "BaseBdev1", 00:14:45.652 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:45.652 "is_configured": false, 00:14:45.652 "data_offset": 0, 00:14:45.652 "data_size": 0 00:14:45.652 }, 00:14:45.652 { 00:14:45.652 "name": "BaseBdev2", 00:14:45.652 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:45.652 "is_configured": false, 00:14:45.652 "data_offset": 0, 00:14:45.652 "data_size": 0 00:14:45.652 }, 00:14:45.652 { 00:14:45.652 "name": "BaseBdev3", 00:14:45.652 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:45.652 "is_configured": false, 00:14:45.653 "data_offset": 0, 00:14:45.653 "data_size": 0 00:14:45.653 } 00:14:45.653 ] 00:14:45.653 }' 00:14:45.653 16:02:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:45.653 16:02:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.913 [2024-11-20 16:02:44.037573] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:45.913 [2024-11-20 16:02:44.037606] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.913 [2024-11-20 16:02:44.045570] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:45.913 [2024-11-20 16:02:44.045611] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:45.913 [2024-11-20 16:02:44.045619] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:45.913 [2024-11-20 16:02:44.045628] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:45.913 [2024-11-20 16:02:44.045634] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:45.913 [2024-11-20 16:02:44.045643] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.913 [2024-11-20 16:02:44.078200] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:45.913 BaseBdev1 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.913 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.914 [ 00:14:45.914 { 00:14:45.914 "name": "BaseBdev1", 00:14:45.914 "aliases": [ 00:14:45.914 "afa0ae48-1ab3-418d-b70e-67b7beb8d976" 00:14:45.914 ], 00:14:45.914 "product_name": "Malloc disk", 00:14:45.914 "block_size": 512, 00:14:45.914 "num_blocks": 65536, 00:14:45.914 "uuid": "afa0ae48-1ab3-418d-b70e-67b7beb8d976", 00:14:45.914 "assigned_rate_limits": { 00:14:45.914 "rw_ios_per_sec": 0, 00:14:45.914 "rw_mbytes_per_sec": 0, 00:14:45.914 "r_mbytes_per_sec": 0, 00:14:45.914 "w_mbytes_per_sec": 0 00:14:45.914 }, 00:14:45.914 "claimed": true, 00:14:45.914 "claim_type": "exclusive_write", 00:14:45.914 "zoned": false, 00:14:45.914 "supported_io_types": { 00:14:45.914 "read": true, 00:14:45.914 "write": true, 00:14:45.914 "unmap": true, 00:14:45.914 "flush": true, 00:14:45.914 "reset": true, 00:14:45.914 "nvme_admin": false, 00:14:45.914 "nvme_io": false, 00:14:45.914 "nvme_io_md": false, 00:14:45.914 "write_zeroes": true, 00:14:45.914 "zcopy": true, 00:14:45.914 "get_zone_info": false, 00:14:45.914 "zone_management": false, 00:14:45.914 "zone_append": false, 00:14:45.914 "compare": false, 00:14:45.914 "compare_and_write": false, 00:14:45.914 "abort": true, 00:14:45.914 "seek_hole": false, 00:14:45.914 "seek_data": false, 00:14:45.914 "copy": true, 00:14:45.914 "nvme_iov_md": false 00:14:45.914 }, 00:14:45.914 "memory_domains": [ 00:14:45.914 { 00:14:45.914 "dma_device_id": "system", 00:14:45.914 "dma_device_type": 1 00:14:45.914 }, 00:14:45.914 { 00:14:45.914 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:45.914 "dma_device_type": 2 00:14:45.914 } 00:14:45.914 ], 00:14:45.914 "driver_specific": {} 00:14:45.914 } 00:14:45.914 ] 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:45.914 "name": "Existed_Raid", 00:14:45.914 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:45.914 "strip_size_kb": 64, 00:14:45.914 "state": "configuring", 00:14:45.914 "raid_level": "raid5f", 00:14:45.914 "superblock": false, 00:14:45.914 "num_base_bdevs": 3, 00:14:45.914 "num_base_bdevs_discovered": 1, 00:14:45.914 "num_base_bdevs_operational": 3, 00:14:45.914 "base_bdevs_list": [ 00:14:45.914 { 00:14:45.914 "name": "BaseBdev1", 00:14:45.914 "uuid": "afa0ae48-1ab3-418d-b70e-67b7beb8d976", 00:14:45.914 "is_configured": true, 00:14:45.914 "data_offset": 0, 00:14:45.914 "data_size": 65536 00:14:45.914 }, 00:14:45.914 { 00:14:45.914 "name": "BaseBdev2", 00:14:45.914 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:45.914 "is_configured": false, 00:14:45.914 "data_offset": 0, 00:14:45.914 "data_size": 0 00:14:45.914 }, 00:14:45.914 { 00:14:45.914 "name": "BaseBdev3", 00:14:45.914 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:45.914 "is_configured": false, 00:14:45.914 "data_offset": 0, 00:14:45.914 "data_size": 0 00:14:45.914 } 00:14:45.914 ] 00:14:45.914 }' 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:45.914 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.174 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:46.174 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.174 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.174 [2024-11-20 16:02:44.422323] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:46.435 [2024-11-20 16:02:44.423822] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.435 [2024-11-20 16:02:44.430381] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:46.435 [2024-11-20 16:02:44.432254] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:46.435 [2024-11-20 16:02:44.432296] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:46.435 [2024-11-20 16:02:44.432306] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:46.435 [2024-11-20 16:02:44.432317] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:46.435 "name": "Existed_Raid", 00:14:46.435 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.435 "strip_size_kb": 64, 00:14:46.435 "state": "configuring", 00:14:46.435 "raid_level": "raid5f", 00:14:46.435 "superblock": false, 00:14:46.435 "num_base_bdevs": 3, 00:14:46.435 "num_base_bdevs_discovered": 1, 00:14:46.435 "num_base_bdevs_operational": 3, 00:14:46.435 "base_bdevs_list": [ 00:14:46.435 { 00:14:46.435 "name": "BaseBdev1", 00:14:46.435 "uuid": "afa0ae48-1ab3-418d-b70e-67b7beb8d976", 00:14:46.435 "is_configured": true, 00:14:46.435 "data_offset": 0, 00:14:46.435 "data_size": 65536 00:14:46.435 }, 00:14:46.435 { 00:14:46.435 "name": "BaseBdev2", 00:14:46.435 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.435 "is_configured": false, 00:14:46.435 "data_offset": 0, 00:14:46.435 "data_size": 0 00:14:46.435 }, 00:14:46.435 { 00:14:46.435 "name": "BaseBdev3", 00:14:46.435 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.435 "is_configured": false, 00:14:46.435 "data_offset": 0, 00:14:46.435 "data_size": 0 00:14:46.435 } 00:14:46.435 ] 00:14:46.435 }' 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:46.435 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.696 [2024-11-20 16:02:44.785017] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:46.696 BaseBdev2 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.696 [ 00:14:46.696 { 00:14:46.696 "name": "BaseBdev2", 00:14:46.696 "aliases": [ 00:14:46.696 "60c023bf-2873-49ac-a725-886542469fe4" 00:14:46.696 ], 00:14:46.696 "product_name": "Malloc disk", 00:14:46.696 "block_size": 512, 00:14:46.696 "num_blocks": 65536, 00:14:46.696 "uuid": "60c023bf-2873-49ac-a725-886542469fe4", 00:14:46.696 "assigned_rate_limits": { 00:14:46.696 "rw_ios_per_sec": 0, 00:14:46.696 "rw_mbytes_per_sec": 0, 00:14:46.696 "r_mbytes_per_sec": 0, 00:14:46.696 "w_mbytes_per_sec": 0 00:14:46.696 }, 00:14:46.696 "claimed": true, 00:14:46.696 "claim_type": "exclusive_write", 00:14:46.696 "zoned": false, 00:14:46.696 "supported_io_types": { 00:14:46.696 "read": true, 00:14:46.696 "write": true, 00:14:46.696 "unmap": true, 00:14:46.696 "flush": true, 00:14:46.696 "reset": true, 00:14:46.696 "nvme_admin": false, 00:14:46.696 "nvme_io": false, 00:14:46.696 "nvme_io_md": false, 00:14:46.696 "write_zeroes": true, 00:14:46.696 "zcopy": true, 00:14:46.696 "get_zone_info": false, 00:14:46.696 "zone_management": false, 00:14:46.696 "zone_append": false, 00:14:46.696 "compare": false, 00:14:46.696 "compare_and_write": false, 00:14:46.696 "abort": true, 00:14:46.696 "seek_hole": false, 00:14:46.696 "seek_data": false, 00:14:46.696 "copy": true, 00:14:46.696 "nvme_iov_md": false 00:14:46.696 }, 00:14:46.696 "memory_domains": [ 00:14:46.696 { 00:14:46.696 "dma_device_id": "system", 00:14:46.696 "dma_device_type": 1 00:14:46.696 }, 00:14:46.696 { 00:14:46.696 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:46.696 "dma_device_type": 2 00:14:46.696 } 00:14:46.696 ], 00:14:46.696 "driver_specific": {} 00:14:46.696 } 00:14:46.696 ] 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.696 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:46.696 "name": "Existed_Raid", 00:14:46.696 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.696 "strip_size_kb": 64, 00:14:46.696 "state": "configuring", 00:14:46.696 "raid_level": "raid5f", 00:14:46.696 "superblock": false, 00:14:46.696 "num_base_bdevs": 3, 00:14:46.696 "num_base_bdevs_discovered": 2, 00:14:46.696 "num_base_bdevs_operational": 3, 00:14:46.696 "base_bdevs_list": [ 00:14:46.696 { 00:14:46.696 "name": "BaseBdev1", 00:14:46.696 "uuid": "afa0ae48-1ab3-418d-b70e-67b7beb8d976", 00:14:46.696 "is_configured": true, 00:14:46.696 "data_offset": 0, 00:14:46.696 "data_size": 65536 00:14:46.696 }, 00:14:46.696 { 00:14:46.696 "name": "BaseBdev2", 00:14:46.696 "uuid": "60c023bf-2873-49ac-a725-886542469fe4", 00:14:46.696 "is_configured": true, 00:14:46.696 "data_offset": 0, 00:14:46.696 "data_size": 65536 00:14:46.696 }, 00:14:46.696 { 00:14:46.696 "name": "BaseBdev3", 00:14:46.696 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.696 "is_configured": false, 00:14:46.696 "data_offset": 0, 00:14:46.696 "data_size": 0 00:14:46.696 } 00:14:46.696 ] 00:14:46.697 }' 00:14:46.697 16:02:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:46.697 16:02:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.956 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:46.956 BaseBdev3 00:14:46.956 [ 00:14:46.956 { 00:14:46.956 "name": "BaseBdev3", 00:14:46.956 "aliases": [ 00:14:46.956 "574d1618-0cd5-4eae-8cc1-b25e2103a75a" 00:14:46.956 ], 00:14:46.956 "product_name": "Malloc disk", 00:14:46.956 "block_size": 512, 00:14:46.956 "num_blocks": 65536, 00:14:46.956 "uuid": "574d1618-0cd5-4eae-8cc1-b25e2103a75a", 00:14:46.956 "assigned_rate_limits": { 00:14:46.956 "rw_ios_per_sec": 0, 00:14:46.956 "rw_mbytes_per_sec": 0, 00:14:46.956 "r_mbytes_per_sec": 0, 00:14:46.956 "w_mbytes_per_sec": 0 00:14:46.956 }, 00:14:46.956 "claimed": true, 00:14:46.956 "claim_type": "exclusive_write", 00:14:46.956 "zoned": false, 00:14:46.956 "supported_io_types": { 00:14:46.956 "read": true, 00:14:46.956 "write": true, 00:14:46.956 "unmap": true, 00:14:46.956 "flush": true, 00:14:46.956 "reset": true, 00:14:46.956 "nvme_admin": false, 00:14:46.956 "nvme_io": false, 00:14:46.956 "nvme_io_md": false, 00:14:46.956 "write_zeroes": true, 00:14:46.956 "zcopy": true, 00:14:46.956 "get_zone_info": false, 00:14:46.956 "zone_management": false, 00:14:46.956 "zone_append": false, 00:14:46.956 "compare": false, 00:14:46.956 "compare_and_write": false, 00:14:46.956 "abort": true, 00:14:46.956 "seek_hole": false, 00:14:46.956 "seek_data": false, 00:14:46.956 "copy": true, 00:14:46.956 "nvme_iov_md": false 00:14:46.956 }, 00:14:46.956 "memory_domains": [ 00:14:46.956 { 00:14:46.956 "dma_device_id": "system", 00:14:46.956 "dma_device_type": 1 00:14:46.956 }, 00:14:46.956 { 00:14:46.956 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:46.956 "dma_device_type": 2 00:14:46.956 } 00:14:46.956 ], 00:14:46.956 "driver_specific": {} 00:14:46.956 } 00:14:46.957 ] 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.957 [2024-11-20 16:02:45.155148] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:46.957 [2024-11-20 16:02:45.155324] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:14:46.957 [2024-11-20 16:02:45.155361] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:14:46.957 [2024-11-20 16:02:45.155664] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:14:46.957 [2024-11-20 16:02:45.159492] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:14:46.957 [2024-11-20 16:02:45.159590] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:14:46.957 [2024-11-20 16:02:45.159908] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.957 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.217 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:47.217 "name": "Existed_Raid", 00:14:47.217 "uuid": "e6b10775-5433-4805-870c-7f6a520791a9", 00:14:47.217 "strip_size_kb": 64, 00:14:47.217 "state": "online", 00:14:47.217 "raid_level": "raid5f", 00:14:47.217 "superblock": false, 00:14:47.217 "num_base_bdevs": 3, 00:14:47.217 "num_base_bdevs_discovered": 3, 00:14:47.217 "num_base_bdevs_operational": 3, 00:14:47.217 "base_bdevs_list": [ 00:14:47.217 { 00:14:47.217 "name": "BaseBdev1", 00:14:47.217 "uuid": "afa0ae48-1ab3-418d-b70e-67b7beb8d976", 00:14:47.217 "is_configured": true, 00:14:47.217 "data_offset": 0, 00:14:47.217 "data_size": 65536 00:14:47.217 }, 00:14:47.217 { 00:14:47.217 "name": "BaseBdev2", 00:14:47.217 "uuid": "60c023bf-2873-49ac-a725-886542469fe4", 00:14:47.217 "is_configured": true, 00:14:47.217 "data_offset": 0, 00:14:47.217 "data_size": 65536 00:14:47.217 }, 00:14:47.217 { 00:14:47.217 "name": "BaseBdev3", 00:14:47.217 "uuid": "574d1618-0cd5-4eae-8cc1-b25e2103a75a", 00:14:47.217 "is_configured": true, 00:14:47.217 "data_offset": 0, 00:14:47.217 "data_size": 65536 00:14:47.217 } 00:14:47.217 ] 00:14:47.217 }' 00:14:47.217 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:47.217 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.477 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:47.477 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:47.477 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:47.477 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:47.477 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:47.477 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:47.477 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:47.477 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:47.477 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.477 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.477 [2024-11-20 16:02:45.512772] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:47.477 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.477 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:47.477 "name": "Existed_Raid", 00:14:47.477 "aliases": [ 00:14:47.477 "e6b10775-5433-4805-870c-7f6a520791a9" 00:14:47.477 ], 00:14:47.477 "product_name": "Raid Volume", 00:14:47.477 "block_size": 512, 00:14:47.477 "num_blocks": 131072, 00:14:47.477 "uuid": "e6b10775-5433-4805-870c-7f6a520791a9", 00:14:47.477 "assigned_rate_limits": { 00:14:47.477 "rw_ios_per_sec": 0, 00:14:47.477 "rw_mbytes_per_sec": 0, 00:14:47.478 "r_mbytes_per_sec": 0, 00:14:47.478 "w_mbytes_per_sec": 0 00:14:47.478 }, 00:14:47.478 "claimed": false, 00:14:47.478 "zoned": false, 00:14:47.478 "supported_io_types": { 00:14:47.478 "read": true, 00:14:47.478 "write": true, 00:14:47.478 "unmap": false, 00:14:47.478 "flush": false, 00:14:47.478 "reset": true, 00:14:47.478 "nvme_admin": false, 00:14:47.478 "nvme_io": false, 00:14:47.478 "nvme_io_md": false, 00:14:47.478 "write_zeroes": true, 00:14:47.478 "zcopy": false, 00:14:47.478 "get_zone_info": false, 00:14:47.478 "zone_management": false, 00:14:47.478 "zone_append": false, 00:14:47.478 "compare": false, 00:14:47.478 "compare_and_write": false, 00:14:47.478 "abort": false, 00:14:47.478 "seek_hole": false, 00:14:47.478 "seek_data": false, 00:14:47.478 "copy": false, 00:14:47.478 "nvme_iov_md": false 00:14:47.478 }, 00:14:47.478 "driver_specific": { 00:14:47.478 "raid": { 00:14:47.478 "uuid": "e6b10775-5433-4805-870c-7f6a520791a9", 00:14:47.478 "strip_size_kb": 64, 00:14:47.478 "state": "online", 00:14:47.478 "raid_level": "raid5f", 00:14:47.478 "superblock": false, 00:14:47.478 "num_base_bdevs": 3, 00:14:47.478 "num_base_bdevs_discovered": 3, 00:14:47.478 "num_base_bdevs_operational": 3, 00:14:47.478 "base_bdevs_list": [ 00:14:47.478 { 00:14:47.478 "name": "BaseBdev1", 00:14:47.478 "uuid": "afa0ae48-1ab3-418d-b70e-67b7beb8d976", 00:14:47.478 "is_configured": true, 00:14:47.478 "data_offset": 0, 00:14:47.478 "data_size": 65536 00:14:47.478 }, 00:14:47.478 { 00:14:47.478 "name": "BaseBdev2", 00:14:47.478 "uuid": "60c023bf-2873-49ac-a725-886542469fe4", 00:14:47.478 "is_configured": true, 00:14:47.478 "data_offset": 0, 00:14:47.478 "data_size": 65536 00:14:47.478 }, 00:14:47.478 { 00:14:47.478 "name": "BaseBdev3", 00:14:47.478 "uuid": "574d1618-0cd5-4eae-8cc1-b25e2103a75a", 00:14:47.478 "is_configured": true, 00:14:47.478 "data_offset": 0, 00:14:47.478 "data_size": 65536 00:14:47.478 } 00:14:47.478 ] 00:14:47.478 } 00:14:47.478 } 00:14:47.478 }' 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:47.478 BaseBdev2 00:14:47.478 BaseBdev3' 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.478 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.478 [2024-11-20 16:02:45.696616] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:47.738 "name": "Existed_Raid", 00:14:47.738 "uuid": "e6b10775-5433-4805-870c-7f6a520791a9", 00:14:47.738 "strip_size_kb": 64, 00:14:47.738 "state": "online", 00:14:47.738 "raid_level": "raid5f", 00:14:47.738 "superblock": false, 00:14:47.738 "num_base_bdevs": 3, 00:14:47.738 "num_base_bdevs_discovered": 2, 00:14:47.738 "num_base_bdevs_operational": 2, 00:14:47.738 "base_bdevs_list": [ 00:14:47.738 { 00:14:47.738 "name": null, 00:14:47.738 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:47.738 "is_configured": false, 00:14:47.738 "data_offset": 0, 00:14:47.738 "data_size": 65536 00:14:47.738 }, 00:14:47.738 { 00:14:47.738 "name": "BaseBdev2", 00:14:47.738 "uuid": "60c023bf-2873-49ac-a725-886542469fe4", 00:14:47.738 "is_configured": true, 00:14:47.738 "data_offset": 0, 00:14:47.738 "data_size": 65536 00:14:47.738 }, 00:14:47.738 { 00:14:47.738 "name": "BaseBdev3", 00:14:47.738 "uuid": "574d1618-0cd5-4eae-8cc1-b25e2103a75a", 00:14:47.738 "is_configured": true, 00:14:47.738 "data_offset": 0, 00:14:47.738 "data_size": 65536 00:14:47.738 } 00:14:47.738 ] 00:14:47.738 }' 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:47.738 16:02:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.997 [2024-11-20 16:02:46.123412] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:47.997 [2024-11-20 16:02:46.123502] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:47.997 [2024-11-20 16:02:46.181435] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.997 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.997 [2024-11-20 16:02:46.221502] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:47.997 [2024-11-20 16:02:46.221647] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.259 BaseBdev2 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.259 [ 00:14:48.259 { 00:14:48.259 "name": "BaseBdev2", 00:14:48.259 "aliases": [ 00:14:48.259 "a2cbb59f-6259-4bcf-9a48-e31473de8686" 00:14:48.259 ], 00:14:48.259 "product_name": "Malloc disk", 00:14:48.259 "block_size": 512, 00:14:48.259 "num_blocks": 65536, 00:14:48.259 "uuid": "a2cbb59f-6259-4bcf-9a48-e31473de8686", 00:14:48.259 "assigned_rate_limits": { 00:14:48.259 "rw_ios_per_sec": 0, 00:14:48.259 "rw_mbytes_per_sec": 0, 00:14:48.259 "r_mbytes_per_sec": 0, 00:14:48.259 "w_mbytes_per_sec": 0 00:14:48.259 }, 00:14:48.259 "claimed": false, 00:14:48.259 "zoned": false, 00:14:48.259 "supported_io_types": { 00:14:48.259 "read": true, 00:14:48.259 "write": true, 00:14:48.259 "unmap": true, 00:14:48.259 "flush": true, 00:14:48.259 "reset": true, 00:14:48.259 "nvme_admin": false, 00:14:48.259 "nvme_io": false, 00:14:48.259 "nvme_io_md": false, 00:14:48.259 "write_zeroes": true, 00:14:48.259 "zcopy": true, 00:14:48.259 "get_zone_info": false, 00:14:48.259 "zone_management": false, 00:14:48.259 "zone_append": false, 00:14:48.259 "compare": false, 00:14:48.259 "compare_and_write": false, 00:14:48.259 "abort": true, 00:14:48.259 "seek_hole": false, 00:14:48.259 "seek_data": false, 00:14:48.259 "copy": true, 00:14:48.259 "nvme_iov_md": false 00:14:48.259 }, 00:14:48.259 "memory_domains": [ 00:14:48.259 { 00:14:48.259 "dma_device_id": "system", 00:14:48.259 "dma_device_type": 1 00:14:48.259 }, 00:14:48.259 { 00:14:48.259 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:48.259 "dma_device_type": 2 00:14:48.259 } 00:14:48.259 ], 00:14:48.259 "driver_specific": {} 00:14:48.259 } 00:14:48.259 ] 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.259 BaseBdev3 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.259 [ 00:14:48.259 { 00:14:48.259 "name": "BaseBdev3", 00:14:48.259 "aliases": [ 00:14:48.259 "90172f3a-504f-49ff-b7cd-1efa426b3655" 00:14:48.259 ], 00:14:48.259 "product_name": "Malloc disk", 00:14:48.259 "block_size": 512, 00:14:48.259 "num_blocks": 65536, 00:14:48.259 "uuid": "90172f3a-504f-49ff-b7cd-1efa426b3655", 00:14:48.259 "assigned_rate_limits": { 00:14:48.259 "rw_ios_per_sec": 0, 00:14:48.259 "rw_mbytes_per_sec": 0, 00:14:48.259 "r_mbytes_per_sec": 0, 00:14:48.259 "w_mbytes_per_sec": 0 00:14:48.259 }, 00:14:48.259 "claimed": false, 00:14:48.259 "zoned": false, 00:14:48.259 "supported_io_types": { 00:14:48.259 "read": true, 00:14:48.259 "write": true, 00:14:48.259 "unmap": true, 00:14:48.259 "flush": true, 00:14:48.259 "reset": true, 00:14:48.259 "nvme_admin": false, 00:14:48.259 "nvme_io": false, 00:14:48.259 "nvme_io_md": false, 00:14:48.259 "write_zeroes": true, 00:14:48.259 "zcopy": true, 00:14:48.259 "get_zone_info": false, 00:14:48.259 "zone_management": false, 00:14:48.259 "zone_append": false, 00:14:48.259 "compare": false, 00:14:48.259 "compare_and_write": false, 00:14:48.259 "abort": true, 00:14:48.259 "seek_hole": false, 00:14:48.259 "seek_data": false, 00:14:48.259 "copy": true, 00:14:48.259 "nvme_iov_md": false 00:14:48.259 }, 00:14:48.259 "memory_domains": [ 00:14:48.259 { 00:14:48.259 "dma_device_id": "system", 00:14:48.259 "dma_device_type": 1 00:14:48.259 }, 00:14:48.259 { 00:14:48.259 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:48.259 "dma_device_type": 2 00:14:48.259 } 00:14:48.259 ], 00:14:48.259 "driver_specific": {} 00:14:48.259 } 00:14:48.259 ] 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.259 [2024-11-20 16:02:46.438091] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:48.259 [2024-11-20 16:02:46.438228] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:48.259 [2024-11-20 16:02:46.438298] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:48.259 [2024-11-20 16:02:46.440182] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:48.259 "name": "Existed_Raid", 00:14:48.259 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:48.259 "strip_size_kb": 64, 00:14:48.259 "state": "configuring", 00:14:48.259 "raid_level": "raid5f", 00:14:48.259 "superblock": false, 00:14:48.259 "num_base_bdevs": 3, 00:14:48.259 "num_base_bdevs_discovered": 2, 00:14:48.259 "num_base_bdevs_operational": 3, 00:14:48.259 "base_bdevs_list": [ 00:14:48.259 { 00:14:48.259 "name": "BaseBdev1", 00:14:48.259 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:48.259 "is_configured": false, 00:14:48.259 "data_offset": 0, 00:14:48.259 "data_size": 0 00:14:48.259 }, 00:14:48.259 { 00:14:48.259 "name": "BaseBdev2", 00:14:48.259 "uuid": "a2cbb59f-6259-4bcf-9a48-e31473de8686", 00:14:48.259 "is_configured": true, 00:14:48.259 "data_offset": 0, 00:14:48.259 "data_size": 65536 00:14:48.259 }, 00:14:48.259 { 00:14:48.259 "name": "BaseBdev3", 00:14:48.259 "uuid": "90172f3a-504f-49ff-b7cd-1efa426b3655", 00:14:48.259 "is_configured": true, 00:14:48.259 "data_offset": 0, 00:14:48.259 "data_size": 65536 00:14:48.259 } 00:14:48.259 ] 00:14:48.259 }' 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:48.259 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.520 [2024-11-20 16:02:46.758177] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.520 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.782 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.782 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:48.782 "name": "Existed_Raid", 00:14:48.782 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:48.782 "strip_size_kb": 64, 00:14:48.782 "state": "configuring", 00:14:48.782 "raid_level": "raid5f", 00:14:48.782 "superblock": false, 00:14:48.782 "num_base_bdevs": 3, 00:14:48.782 "num_base_bdevs_discovered": 1, 00:14:48.782 "num_base_bdevs_operational": 3, 00:14:48.782 "base_bdevs_list": [ 00:14:48.782 { 00:14:48.782 "name": "BaseBdev1", 00:14:48.782 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:48.782 "is_configured": false, 00:14:48.782 "data_offset": 0, 00:14:48.782 "data_size": 0 00:14:48.782 }, 00:14:48.782 { 00:14:48.782 "name": null, 00:14:48.782 "uuid": "a2cbb59f-6259-4bcf-9a48-e31473de8686", 00:14:48.782 "is_configured": false, 00:14:48.782 "data_offset": 0, 00:14:48.782 "data_size": 65536 00:14:48.782 }, 00:14:48.782 { 00:14:48.782 "name": "BaseBdev3", 00:14:48.782 "uuid": "90172f3a-504f-49ff-b7cd-1efa426b3655", 00:14:48.782 "is_configured": true, 00:14:48.782 "data_offset": 0, 00:14:48.782 "data_size": 65536 00:14:48.782 } 00:14:48.782 ] 00:14:48.782 }' 00:14:48.782 16:02:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:48.782 16:02:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.043 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.043 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.043 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.044 [2024-11-20 16:02:47.125019] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:49.044 BaseBdev1 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.044 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.044 [ 00:14:49.044 { 00:14:49.044 "name": "BaseBdev1", 00:14:49.044 "aliases": [ 00:14:49.044 "d6605fd2-12b0-46a6-bf5a-3a7383198944" 00:14:49.044 ], 00:14:49.044 "product_name": "Malloc disk", 00:14:49.044 "block_size": 512, 00:14:49.045 "num_blocks": 65536, 00:14:49.045 "uuid": "d6605fd2-12b0-46a6-bf5a-3a7383198944", 00:14:49.045 "assigned_rate_limits": { 00:14:49.045 "rw_ios_per_sec": 0, 00:14:49.045 "rw_mbytes_per_sec": 0, 00:14:49.045 "r_mbytes_per_sec": 0, 00:14:49.045 "w_mbytes_per_sec": 0 00:14:49.045 }, 00:14:49.045 "claimed": true, 00:14:49.045 "claim_type": "exclusive_write", 00:14:49.045 "zoned": false, 00:14:49.045 "supported_io_types": { 00:14:49.045 "read": true, 00:14:49.045 "write": true, 00:14:49.045 "unmap": true, 00:14:49.045 "flush": true, 00:14:49.045 "reset": true, 00:14:49.045 "nvme_admin": false, 00:14:49.045 "nvme_io": false, 00:14:49.045 "nvme_io_md": false, 00:14:49.045 "write_zeroes": true, 00:14:49.045 "zcopy": true, 00:14:49.045 "get_zone_info": false, 00:14:49.045 "zone_management": false, 00:14:49.045 "zone_append": false, 00:14:49.045 "compare": false, 00:14:49.045 "compare_and_write": false, 00:14:49.045 "abort": true, 00:14:49.045 "seek_hole": false, 00:14:49.045 "seek_data": false, 00:14:49.045 "copy": true, 00:14:49.045 "nvme_iov_md": false 00:14:49.045 }, 00:14:49.045 "memory_domains": [ 00:14:49.045 { 00:14:49.045 "dma_device_id": "system", 00:14:49.045 "dma_device_type": 1 00:14:49.045 }, 00:14:49.046 { 00:14:49.046 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:49.046 "dma_device_type": 2 00:14:49.046 } 00:14:49.046 ], 00:14:49.046 "driver_specific": {} 00:14:49.046 } 00:14:49.046 ] 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.046 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:49.046 "name": "Existed_Raid", 00:14:49.046 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:49.046 "strip_size_kb": 64, 00:14:49.046 "state": "configuring", 00:14:49.046 "raid_level": "raid5f", 00:14:49.046 "superblock": false, 00:14:49.046 "num_base_bdevs": 3, 00:14:49.046 "num_base_bdevs_discovered": 2, 00:14:49.046 "num_base_bdevs_operational": 3, 00:14:49.046 "base_bdevs_list": [ 00:14:49.046 { 00:14:49.046 "name": "BaseBdev1", 00:14:49.046 "uuid": "d6605fd2-12b0-46a6-bf5a-3a7383198944", 00:14:49.046 "is_configured": true, 00:14:49.046 "data_offset": 0, 00:14:49.046 "data_size": 65536 00:14:49.046 }, 00:14:49.047 { 00:14:49.047 "name": null, 00:14:49.047 "uuid": "a2cbb59f-6259-4bcf-9a48-e31473de8686", 00:14:49.047 "is_configured": false, 00:14:49.047 "data_offset": 0, 00:14:49.047 "data_size": 65536 00:14:49.047 }, 00:14:49.047 { 00:14:49.047 "name": "BaseBdev3", 00:14:49.047 "uuid": "90172f3a-504f-49ff-b7cd-1efa426b3655", 00:14:49.047 "is_configured": true, 00:14:49.047 "data_offset": 0, 00:14:49.047 "data_size": 65536 00:14:49.047 } 00:14:49.047 ] 00:14:49.047 }' 00:14:49.047 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:49.047 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.310 [2024-11-20 16:02:47.501159] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:49.310 "name": "Existed_Raid", 00:14:49.310 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:49.310 "strip_size_kb": 64, 00:14:49.310 "state": "configuring", 00:14:49.310 "raid_level": "raid5f", 00:14:49.310 "superblock": false, 00:14:49.310 "num_base_bdevs": 3, 00:14:49.310 "num_base_bdevs_discovered": 1, 00:14:49.310 "num_base_bdevs_operational": 3, 00:14:49.310 "base_bdevs_list": [ 00:14:49.310 { 00:14:49.310 "name": "BaseBdev1", 00:14:49.310 "uuid": "d6605fd2-12b0-46a6-bf5a-3a7383198944", 00:14:49.310 "is_configured": true, 00:14:49.310 "data_offset": 0, 00:14:49.310 "data_size": 65536 00:14:49.310 }, 00:14:49.310 { 00:14:49.310 "name": null, 00:14:49.310 "uuid": "a2cbb59f-6259-4bcf-9a48-e31473de8686", 00:14:49.310 "is_configured": false, 00:14:49.310 "data_offset": 0, 00:14:49.310 "data_size": 65536 00:14:49.310 }, 00:14:49.310 { 00:14:49.310 "name": null, 00:14:49.310 "uuid": "90172f3a-504f-49ff-b7cd-1efa426b3655", 00:14:49.310 "is_configured": false, 00:14:49.310 "data_offset": 0, 00:14:49.310 "data_size": 65536 00:14:49.310 } 00:14:49.310 ] 00:14:49.310 }' 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:49.310 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.731 [2024-11-20 16:02:47.865294] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:49.731 "name": "Existed_Raid", 00:14:49.731 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:49.731 "strip_size_kb": 64, 00:14:49.731 "state": "configuring", 00:14:49.731 "raid_level": "raid5f", 00:14:49.731 "superblock": false, 00:14:49.731 "num_base_bdevs": 3, 00:14:49.731 "num_base_bdevs_discovered": 2, 00:14:49.731 "num_base_bdevs_operational": 3, 00:14:49.731 "base_bdevs_list": [ 00:14:49.731 { 00:14:49.731 "name": "BaseBdev1", 00:14:49.731 "uuid": "d6605fd2-12b0-46a6-bf5a-3a7383198944", 00:14:49.731 "is_configured": true, 00:14:49.731 "data_offset": 0, 00:14:49.731 "data_size": 65536 00:14:49.731 }, 00:14:49.731 { 00:14:49.731 "name": null, 00:14:49.731 "uuid": "a2cbb59f-6259-4bcf-9a48-e31473de8686", 00:14:49.731 "is_configured": false, 00:14:49.731 "data_offset": 0, 00:14:49.731 "data_size": 65536 00:14:49.731 }, 00:14:49.731 { 00:14:49.731 "name": "BaseBdev3", 00:14:49.731 "uuid": "90172f3a-504f-49ff-b7cd-1efa426b3655", 00:14:49.731 "is_configured": true, 00:14:49.731 "data_offset": 0, 00:14:49.731 "data_size": 65536 00:14:49.731 } 00:14:49.731 ] 00:14:49.731 }' 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:49.731 16:02:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.992 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.992 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.992 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:49.992 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.992 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.992 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:49.992 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:49.992 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.992 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.992 [2024-11-20 16:02:48.221393] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:50.253 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.253 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:50.254 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:50.254 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:50.254 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:50.254 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:50.254 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:50.254 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:50.254 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:50.254 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:50.254 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:50.254 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.254 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:50.254 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.254 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.254 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.254 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:50.254 "name": "Existed_Raid", 00:14:50.254 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.254 "strip_size_kb": 64, 00:14:50.254 "state": "configuring", 00:14:50.254 "raid_level": "raid5f", 00:14:50.254 "superblock": false, 00:14:50.254 "num_base_bdevs": 3, 00:14:50.254 "num_base_bdevs_discovered": 1, 00:14:50.254 "num_base_bdevs_operational": 3, 00:14:50.254 "base_bdevs_list": [ 00:14:50.254 { 00:14:50.254 "name": null, 00:14:50.254 "uuid": "d6605fd2-12b0-46a6-bf5a-3a7383198944", 00:14:50.254 "is_configured": false, 00:14:50.254 "data_offset": 0, 00:14:50.254 "data_size": 65536 00:14:50.254 }, 00:14:50.254 { 00:14:50.254 "name": null, 00:14:50.254 "uuid": "a2cbb59f-6259-4bcf-9a48-e31473de8686", 00:14:50.254 "is_configured": false, 00:14:50.254 "data_offset": 0, 00:14:50.254 "data_size": 65536 00:14:50.254 }, 00:14:50.254 { 00:14:50.254 "name": "BaseBdev3", 00:14:50.254 "uuid": "90172f3a-504f-49ff-b7cd-1efa426b3655", 00:14:50.254 "is_configured": true, 00:14:50.254 "data_offset": 0, 00:14:50.254 "data_size": 65536 00:14:50.254 } 00:14:50.254 ] 00:14:50.254 }' 00:14:50.254 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:50.254 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.516 [2024-11-20 16:02:48.632368] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.516 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:50.516 "name": "Existed_Raid", 00:14:50.516 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.516 "strip_size_kb": 64, 00:14:50.516 "state": "configuring", 00:14:50.516 "raid_level": "raid5f", 00:14:50.516 "superblock": false, 00:14:50.516 "num_base_bdevs": 3, 00:14:50.516 "num_base_bdevs_discovered": 2, 00:14:50.516 "num_base_bdevs_operational": 3, 00:14:50.516 "base_bdevs_list": [ 00:14:50.516 { 00:14:50.516 "name": null, 00:14:50.516 "uuid": "d6605fd2-12b0-46a6-bf5a-3a7383198944", 00:14:50.516 "is_configured": false, 00:14:50.516 "data_offset": 0, 00:14:50.517 "data_size": 65536 00:14:50.517 }, 00:14:50.517 { 00:14:50.517 "name": "BaseBdev2", 00:14:50.517 "uuid": "a2cbb59f-6259-4bcf-9a48-e31473de8686", 00:14:50.517 "is_configured": true, 00:14:50.517 "data_offset": 0, 00:14:50.517 "data_size": 65536 00:14:50.517 }, 00:14:50.517 { 00:14:50.517 "name": "BaseBdev3", 00:14:50.517 "uuid": "90172f3a-504f-49ff-b7cd-1efa426b3655", 00:14:50.517 "is_configured": true, 00:14:50.517 "data_offset": 0, 00:14:50.517 "data_size": 65536 00:14:50.517 } 00:14:50.517 ] 00:14:50.517 }' 00:14:50.517 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:50.517 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.778 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.778 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.778 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:50.778 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.778 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.778 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:50.778 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:50.778 16:02:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.778 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.778 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.778 16:02:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.778 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u d6605fd2-12b0-46a6-bf5a-3a7383198944 00:14:50.778 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.778 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.039 [2024-11-20 16:02:49.042969] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:51.039 [2024-11-20 16:02:49.043133] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:14:51.039 [2024-11-20 16:02:49.043151] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:14:51.039 [2024-11-20 16:02:49.043401] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:14:51.039 [2024-11-20 16:02:49.046936] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:14:51.039 [2024-11-20 16:02:49.046954] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:14:51.039 [2024-11-20 16:02:49.047180] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:51.039 NewBaseBdev 00:14:51.039 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.039 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:51.039 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:14:51.039 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:51.039 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:51.039 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:51.039 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:51.039 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:51.039 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.039 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.039 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.039 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:51.039 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.039 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.039 [ 00:14:51.039 { 00:14:51.039 "name": "NewBaseBdev", 00:14:51.039 "aliases": [ 00:14:51.039 "d6605fd2-12b0-46a6-bf5a-3a7383198944" 00:14:51.039 ], 00:14:51.039 "product_name": "Malloc disk", 00:14:51.039 "block_size": 512, 00:14:51.039 "num_blocks": 65536, 00:14:51.039 "uuid": "d6605fd2-12b0-46a6-bf5a-3a7383198944", 00:14:51.039 "assigned_rate_limits": { 00:14:51.039 "rw_ios_per_sec": 0, 00:14:51.039 "rw_mbytes_per_sec": 0, 00:14:51.039 "r_mbytes_per_sec": 0, 00:14:51.039 "w_mbytes_per_sec": 0 00:14:51.039 }, 00:14:51.039 "claimed": true, 00:14:51.039 "claim_type": "exclusive_write", 00:14:51.039 "zoned": false, 00:14:51.039 "supported_io_types": { 00:14:51.039 "read": true, 00:14:51.040 "write": true, 00:14:51.040 "unmap": true, 00:14:51.040 "flush": true, 00:14:51.040 "reset": true, 00:14:51.040 "nvme_admin": false, 00:14:51.040 "nvme_io": false, 00:14:51.040 "nvme_io_md": false, 00:14:51.040 "write_zeroes": true, 00:14:51.040 "zcopy": true, 00:14:51.040 "get_zone_info": false, 00:14:51.040 "zone_management": false, 00:14:51.040 "zone_append": false, 00:14:51.040 "compare": false, 00:14:51.040 "compare_and_write": false, 00:14:51.040 "abort": true, 00:14:51.040 "seek_hole": false, 00:14:51.040 "seek_data": false, 00:14:51.040 "copy": true, 00:14:51.040 "nvme_iov_md": false 00:14:51.040 }, 00:14:51.040 "memory_domains": [ 00:14:51.040 { 00:14:51.040 "dma_device_id": "system", 00:14:51.040 "dma_device_type": 1 00:14:51.040 }, 00:14:51.040 { 00:14:51.040 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:51.040 "dma_device_type": 2 00:14:51.040 } 00:14:51.040 ], 00:14:51.040 "driver_specific": {} 00:14:51.040 } 00:14:51.040 ] 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:51.040 "name": "Existed_Raid", 00:14:51.040 "uuid": "19d912c5-0872-4d26-a860-bf27a01e97dd", 00:14:51.040 "strip_size_kb": 64, 00:14:51.040 "state": "online", 00:14:51.040 "raid_level": "raid5f", 00:14:51.040 "superblock": false, 00:14:51.040 "num_base_bdevs": 3, 00:14:51.040 "num_base_bdevs_discovered": 3, 00:14:51.040 "num_base_bdevs_operational": 3, 00:14:51.040 "base_bdevs_list": [ 00:14:51.040 { 00:14:51.040 "name": "NewBaseBdev", 00:14:51.040 "uuid": "d6605fd2-12b0-46a6-bf5a-3a7383198944", 00:14:51.040 "is_configured": true, 00:14:51.040 "data_offset": 0, 00:14:51.040 "data_size": 65536 00:14:51.040 }, 00:14:51.040 { 00:14:51.040 "name": "BaseBdev2", 00:14:51.040 "uuid": "a2cbb59f-6259-4bcf-9a48-e31473de8686", 00:14:51.040 "is_configured": true, 00:14:51.040 "data_offset": 0, 00:14:51.040 "data_size": 65536 00:14:51.040 }, 00:14:51.040 { 00:14:51.040 "name": "BaseBdev3", 00:14:51.040 "uuid": "90172f3a-504f-49ff-b7cd-1efa426b3655", 00:14:51.040 "is_configured": true, 00:14:51.040 "data_offset": 0, 00:14:51.040 "data_size": 65536 00:14:51.040 } 00:14:51.040 ] 00:14:51.040 }' 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:51.040 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:51.301 [2024-11-20 16:02:49.395432] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:51.301 "name": "Existed_Raid", 00:14:51.301 "aliases": [ 00:14:51.301 "19d912c5-0872-4d26-a860-bf27a01e97dd" 00:14:51.301 ], 00:14:51.301 "product_name": "Raid Volume", 00:14:51.301 "block_size": 512, 00:14:51.301 "num_blocks": 131072, 00:14:51.301 "uuid": "19d912c5-0872-4d26-a860-bf27a01e97dd", 00:14:51.301 "assigned_rate_limits": { 00:14:51.301 "rw_ios_per_sec": 0, 00:14:51.301 "rw_mbytes_per_sec": 0, 00:14:51.301 "r_mbytes_per_sec": 0, 00:14:51.301 "w_mbytes_per_sec": 0 00:14:51.301 }, 00:14:51.301 "claimed": false, 00:14:51.301 "zoned": false, 00:14:51.301 "supported_io_types": { 00:14:51.301 "read": true, 00:14:51.301 "write": true, 00:14:51.301 "unmap": false, 00:14:51.301 "flush": false, 00:14:51.301 "reset": true, 00:14:51.301 "nvme_admin": false, 00:14:51.301 "nvme_io": false, 00:14:51.301 "nvme_io_md": false, 00:14:51.301 "write_zeroes": true, 00:14:51.301 "zcopy": false, 00:14:51.301 "get_zone_info": false, 00:14:51.301 "zone_management": false, 00:14:51.301 "zone_append": false, 00:14:51.301 "compare": false, 00:14:51.301 "compare_and_write": false, 00:14:51.301 "abort": false, 00:14:51.301 "seek_hole": false, 00:14:51.301 "seek_data": false, 00:14:51.301 "copy": false, 00:14:51.301 "nvme_iov_md": false 00:14:51.301 }, 00:14:51.301 "driver_specific": { 00:14:51.301 "raid": { 00:14:51.301 "uuid": "19d912c5-0872-4d26-a860-bf27a01e97dd", 00:14:51.301 "strip_size_kb": 64, 00:14:51.301 "state": "online", 00:14:51.301 "raid_level": "raid5f", 00:14:51.301 "superblock": false, 00:14:51.301 "num_base_bdevs": 3, 00:14:51.301 "num_base_bdevs_discovered": 3, 00:14:51.301 "num_base_bdevs_operational": 3, 00:14:51.301 "base_bdevs_list": [ 00:14:51.301 { 00:14:51.301 "name": "NewBaseBdev", 00:14:51.301 "uuid": "d6605fd2-12b0-46a6-bf5a-3a7383198944", 00:14:51.301 "is_configured": true, 00:14:51.301 "data_offset": 0, 00:14:51.301 "data_size": 65536 00:14:51.301 }, 00:14:51.301 { 00:14:51.301 "name": "BaseBdev2", 00:14:51.301 "uuid": "a2cbb59f-6259-4bcf-9a48-e31473de8686", 00:14:51.301 "is_configured": true, 00:14:51.301 "data_offset": 0, 00:14:51.301 "data_size": 65536 00:14:51.301 }, 00:14:51.301 { 00:14:51.301 "name": "BaseBdev3", 00:14:51.301 "uuid": "90172f3a-504f-49ff-b7cd-1efa426b3655", 00:14:51.301 "is_configured": true, 00:14:51.301 "data_offset": 0, 00:14:51.301 "data_size": 65536 00:14:51.301 } 00:14:51.301 ] 00:14:51.301 } 00:14:51.301 } 00:14:51.301 }' 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:51.301 BaseBdev2 00:14:51.301 BaseBdev3' 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:51.301 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.564 [2024-11-20 16:02:49.591259] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:51.564 [2024-11-20 16:02:49.591282] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:51.564 [2024-11-20 16:02:49.591348] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:51.564 [2024-11-20 16:02:49.591620] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:51.564 [2024-11-20 16:02:49.591632] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 77728 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 77728 ']' 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # kill -0 77728 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # uname 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 77728 00:14:51.564 killing process with pid 77728 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 77728' 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@973 -- # kill 77728 00:14:51.564 [2024-11-20 16:02:49.621493] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:51.564 16:02:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@978 -- # wait 77728 00:14:51.564 [2024-11-20 16:02:49.810581] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:52.506 ************************************ 00:14:52.506 END TEST raid5f_state_function_test 00:14:52.506 ************************************ 00:14:52.506 16:02:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:14:52.506 00:14:52.506 real 0m7.732s 00:14:52.506 user 0m12.278s 00:14:52.506 sys 0m1.273s 00:14:52.506 16:02:50 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:52.506 16:02:50 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.506 16:02:50 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 3 true 00:14:52.506 16:02:50 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:14:52.506 16:02:50 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:52.506 16:02:50 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:52.506 ************************************ 00:14:52.506 START TEST raid5f_state_function_test_sb 00:14:52.507 ************************************ 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 3 true 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:52.507 Process raid pid: 78316 00:14:52.507 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=78316 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 78316' 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 78316 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 78316 ']' 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:52.507 16:02:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:52.507 [2024-11-20 16:02:50.655318] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:14:52.507 [2024-11-20 16:02:50.655467] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:52.808 [2024-11-20 16:02:50.816949] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:52.808 [2024-11-20 16:02:50.919265] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:53.103 [2024-11-20 16:02:51.057519] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:53.103 [2024-11-20 16:02:51.057553] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.363 [2024-11-20 16:02:51.521542] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:53.363 [2024-11-20 16:02:51.521593] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:53.363 [2024-11-20 16:02:51.521609] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:53.363 [2024-11-20 16:02:51.521620] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:53.363 [2024-11-20 16:02:51.521627] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:53.363 [2024-11-20 16:02:51.521637] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:53.363 "name": "Existed_Raid", 00:14:53.363 "uuid": "b86ab67a-d76f-45f0-9793-5d74a76749f1", 00:14:53.363 "strip_size_kb": 64, 00:14:53.363 "state": "configuring", 00:14:53.363 "raid_level": "raid5f", 00:14:53.363 "superblock": true, 00:14:53.363 "num_base_bdevs": 3, 00:14:53.363 "num_base_bdevs_discovered": 0, 00:14:53.363 "num_base_bdevs_operational": 3, 00:14:53.363 "base_bdevs_list": [ 00:14:53.363 { 00:14:53.363 "name": "BaseBdev1", 00:14:53.363 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:53.363 "is_configured": false, 00:14:53.363 "data_offset": 0, 00:14:53.363 "data_size": 0 00:14:53.363 }, 00:14:53.363 { 00:14:53.363 "name": "BaseBdev2", 00:14:53.363 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:53.363 "is_configured": false, 00:14:53.363 "data_offset": 0, 00:14:53.363 "data_size": 0 00:14:53.363 }, 00:14:53.363 { 00:14:53.363 "name": "BaseBdev3", 00:14:53.363 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:53.363 "is_configured": false, 00:14:53.363 "data_offset": 0, 00:14:53.363 "data_size": 0 00:14:53.363 } 00:14:53.363 ] 00:14:53.363 }' 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:53.363 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.624 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:53.625 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.625 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.625 [2024-11-20 16:02:51.841570] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:53.625 [2024-11-20 16:02:51.841604] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:14:53.625 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.625 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:53.625 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.625 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.625 [2024-11-20 16:02:51.849596] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:53.625 [2024-11-20 16:02:51.849642] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:53.625 [2024-11-20 16:02:51.849651] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:53.625 [2024-11-20 16:02:51.849662] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:53.625 [2024-11-20 16:02:51.849678] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:53.625 [2024-11-20 16:02:51.849688] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:53.625 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.625 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:53.625 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.625 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.886 [2024-11-20 16:02:51.882598] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:53.886 BaseBdev1 00:14:53.886 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.886 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:53.886 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:53.886 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:53.886 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:53.886 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:53.886 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:53.886 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:53.886 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.886 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.886 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.886 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:53.886 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.886 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.886 [ 00:14:53.886 { 00:14:53.886 "name": "BaseBdev1", 00:14:53.886 "aliases": [ 00:14:53.886 "5626b12b-413c-41c9-841d-a4bb1cc0e22f" 00:14:53.886 ], 00:14:53.886 "product_name": "Malloc disk", 00:14:53.886 "block_size": 512, 00:14:53.886 "num_blocks": 65536, 00:14:53.887 "uuid": "5626b12b-413c-41c9-841d-a4bb1cc0e22f", 00:14:53.887 "assigned_rate_limits": { 00:14:53.887 "rw_ios_per_sec": 0, 00:14:53.887 "rw_mbytes_per_sec": 0, 00:14:53.887 "r_mbytes_per_sec": 0, 00:14:53.887 "w_mbytes_per_sec": 0 00:14:53.887 }, 00:14:53.887 "claimed": true, 00:14:53.887 "claim_type": "exclusive_write", 00:14:53.887 "zoned": false, 00:14:53.887 "supported_io_types": { 00:14:53.887 "read": true, 00:14:53.887 "write": true, 00:14:53.887 "unmap": true, 00:14:53.887 "flush": true, 00:14:53.887 "reset": true, 00:14:53.887 "nvme_admin": false, 00:14:53.887 "nvme_io": false, 00:14:53.887 "nvme_io_md": false, 00:14:53.887 "write_zeroes": true, 00:14:53.887 "zcopy": true, 00:14:53.887 "get_zone_info": false, 00:14:53.887 "zone_management": false, 00:14:53.887 "zone_append": false, 00:14:53.887 "compare": false, 00:14:53.887 "compare_and_write": false, 00:14:53.887 "abort": true, 00:14:53.887 "seek_hole": false, 00:14:53.887 "seek_data": false, 00:14:53.887 "copy": true, 00:14:53.887 "nvme_iov_md": false 00:14:53.887 }, 00:14:53.887 "memory_domains": [ 00:14:53.887 { 00:14:53.887 "dma_device_id": "system", 00:14:53.887 "dma_device_type": 1 00:14:53.887 }, 00:14:53.887 { 00:14:53.887 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:53.887 "dma_device_type": 2 00:14:53.887 } 00:14:53.887 ], 00:14:53.887 "driver_specific": {} 00:14:53.887 } 00:14:53.887 ] 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:53.887 "name": "Existed_Raid", 00:14:53.887 "uuid": "f2a1c0eb-d6e6-48da-b53c-f8911c5c0c15", 00:14:53.887 "strip_size_kb": 64, 00:14:53.887 "state": "configuring", 00:14:53.887 "raid_level": "raid5f", 00:14:53.887 "superblock": true, 00:14:53.887 "num_base_bdevs": 3, 00:14:53.887 "num_base_bdevs_discovered": 1, 00:14:53.887 "num_base_bdevs_operational": 3, 00:14:53.887 "base_bdevs_list": [ 00:14:53.887 { 00:14:53.887 "name": "BaseBdev1", 00:14:53.887 "uuid": "5626b12b-413c-41c9-841d-a4bb1cc0e22f", 00:14:53.887 "is_configured": true, 00:14:53.887 "data_offset": 2048, 00:14:53.887 "data_size": 63488 00:14:53.887 }, 00:14:53.887 { 00:14:53.887 "name": "BaseBdev2", 00:14:53.887 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:53.887 "is_configured": false, 00:14:53.887 "data_offset": 0, 00:14:53.887 "data_size": 0 00:14:53.887 }, 00:14:53.887 { 00:14:53.887 "name": "BaseBdev3", 00:14:53.887 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:53.887 "is_configured": false, 00:14:53.887 "data_offset": 0, 00:14:53.887 "data_size": 0 00:14:53.887 } 00:14:53.887 ] 00:14:53.887 }' 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:53.887 16:02:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.148 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:54.148 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.148 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.148 [2024-11-20 16:02:52.226811] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:54.148 [2024-11-20 16:02:52.227056] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:14:54.148 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.148 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:54.148 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.148 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.148 [2024-11-20 16:02:52.234895] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:54.148 [2024-11-20 16:02:52.237640] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:54.148 [2024-11-20 16:02:52.237717] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:54.148 [2024-11-20 16:02:52.237736] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:54.148 [2024-11-20 16:02:52.237755] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:54.148 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.148 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:54.148 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:54.148 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:54.148 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:54.148 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:54.148 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:54.148 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:54.148 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:54.148 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:54.149 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:54.149 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:54.149 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:54.149 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.149 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.149 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.149 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:54.149 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.149 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:54.149 "name": "Existed_Raid", 00:14:54.149 "uuid": "fd7d2bdb-760c-4a98-bd49-141cabe5d3da", 00:14:54.149 "strip_size_kb": 64, 00:14:54.149 "state": "configuring", 00:14:54.149 "raid_level": "raid5f", 00:14:54.149 "superblock": true, 00:14:54.149 "num_base_bdevs": 3, 00:14:54.149 "num_base_bdevs_discovered": 1, 00:14:54.149 "num_base_bdevs_operational": 3, 00:14:54.149 "base_bdevs_list": [ 00:14:54.149 { 00:14:54.149 "name": "BaseBdev1", 00:14:54.149 "uuid": "5626b12b-413c-41c9-841d-a4bb1cc0e22f", 00:14:54.149 "is_configured": true, 00:14:54.149 "data_offset": 2048, 00:14:54.149 "data_size": 63488 00:14:54.149 }, 00:14:54.149 { 00:14:54.149 "name": "BaseBdev2", 00:14:54.149 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:54.149 "is_configured": false, 00:14:54.149 "data_offset": 0, 00:14:54.149 "data_size": 0 00:14:54.149 }, 00:14:54.149 { 00:14:54.149 "name": "BaseBdev3", 00:14:54.149 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:54.149 "is_configured": false, 00:14:54.149 "data_offset": 0, 00:14:54.149 "data_size": 0 00:14:54.149 } 00:14:54.149 ] 00:14:54.149 }' 00:14:54.149 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:54.149 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.409 [2024-11-20 16:02:52.577648] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:54.409 BaseBdev2 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.409 [ 00:14:54.409 { 00:14:54.409 "name": "BaseBdev2", 00:14:54.409 "aliases": [ 00:14:54.409 "1f3813f4-0ca5-4421-babc-f56f1666f9a0" 00:14:54.409 ], 00:14:54.409 "product_name": "Malloc disk", 00:14:54.409 "block_size": 512, 00:14:54.409 "num_blocks": 65536, 00:14:54.409 "uuid": "1f3813f4-0ca5-4421-babc-f56f1666f9a0", 00:14:54.409 "assigned_rate_limits": { 00:14:54.409 "rw_ios_per_sec": 0, 00:14:54.409 "rw_mbytes_per_sec": 0, 00:14:54.409 "r_mbytes_per_sec": 0, 00:14:54.409 "w_mbytes_per_sec": 0 00:14:54.409 }, 00:14:54.409 "claimed": true, 00:14:54.409 "claim_type": "exclusive_write", 00:14:54.409 "zoned": false, 00:14:54.409 "supported_io_types": { 00:14:54.409 "read": true, 00:14:54.409 "write": true, 00:14:54.409 "unmap": true, 00:14:54.409 "flush": true, 00:14:54.409 "reset": true, 00:14:54.409 "nvme_admin": false, 00:14:54.409 "nvme_io": false, 00:14:54.409 "nvme_io_md": false, 00:14:54.409 "write_zeroes": true, 00:14:54.409 "zcopy": true, 00:14:54.409 "get_zone_info": false, 00:14:54.409 "zone_management": false, 00:14:54.409 "zone_append": false, 00:14:54.409 "compare": false, 00:14:54.409 "compare_and_write": false, 00:14:54.409 "abort": true, 00:14:54.409 "seek_hole": false, 00:14:54.409 "seek_data": false, 00:14:54.409 "copy": true, 00:14:54.409 "nvme_iov_md": false 00:14:54.409 }, 00:14:54.409 "memory_domains": [ 00:14:54.409 { 00:14:54.409 "dma_device_id": "system", 00:14:54.409 "dma_device_type": 1 00:14:54.409 }, 00:14:54.409 { 00:14:54.409 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:54.409 "dma_device_type": 2 00:14:54.409 } 00:14:54.409 ], 00:14:54.409 "driver_specific": {} 00:14:54.409 } 00:14:54.409 ] 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.409 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:54.409 "name": "Existed_Raid", 00:14:54.409 "uuid": "fd7d2bdb-760c-4a98-bd49-141cabe5d3da", 00:14:54.409 "strip_size_kb": 64, 00:14:54.409 "state": "configuring", 00:14:54.409 "raid_level": "raid5f", 00:14:54.409 "superblock": true, 00:14:54.409 "num_base_bdevs": 3, 00:14:54.409 "num_base_bdevs_discovered": 2, 00:14:54.409 "num_base_bdevs_operational": 3, 00:14:54.409 "base_bdevs_list": [ 00:14:54.409 { 00:14:54.410 "name": "BaseBdev1", 00:14:54.410 "uuid": "5626b12b-413c-41c9-841d-a4bb1cc0e22f", 00:14:54.410 "is_configured": true, 00:14:54.410 "data_offset": 2048, 00:14:54.410 "data_size": 63488 00:14:54.410 }, 00:14:54.410 { 00:14:54.410 "name": "BaseBdev2", 00:14:54.410 "uuid": "1f3813f4-0ca5-4421-babc-f56f1666f9a0", 00:14:54.410 "is_configured": true, 00:14:54.410 "data_offset": 2048, 00:14:54.410 "data_size": 63488 00:14:54.410 }, 00:14:54.410 { 00:14:54.410 "name": "BaseBdev3", 00:14:54.410 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:54.410 "is_configured": false, 00:14:54.410 "data_offset": 0, 00:14:54.410 "data_size": 0 00:14:54.410 } 00:14:54.410 ] 00:14:54.410 }' 00:14:54.410 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:54.410 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.981 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:54.981 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.981 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.981 [2024-11-20 16:02:52.989178] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:54.981 [2024-11-20 16:02:52.989564] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:14:54.981 [2024-11-20 16:02:52.989665] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:54.981 BaseBdev3 00:14:54.981 [2024-11-20 16:02:52.989973] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:14:54.981 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.981 16:02:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:54.981 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:54.981 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:54.981 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:54.981 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:54.981 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:54.981 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:54.981 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.981 16:02:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.981 [2024-11-20 16:02:52.993783] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:14:54.981 [2024-11-20 16:02:52.993798] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:14:54.981 [2024-11-20 16:02:52.993952] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.981 [ 00:14:54.981 { 00:14:54.981 "name": "BaseBdev3", 00:14:54.981 "aliases": [ 00:14:54.981 "8389cd3e-0526-4d8a-8d83-0ac82bfb3ef9" 00:14:54.981 ], 00:14:54.981 "product_name": "Malloc disk", 00:14:54.981 "block_size": 512, 00:14:54.981 "num_blocks": 65536, 00:14:54.981 "uuid": "8389cd3e-0526-4d8a-8d83-0ac82bfb3ef9", 00:14:54.981 "assigned_rate_limits": { 00:14:54.981 "rw_ios_per_sec": 0, 00:14:54.981 "rw_mbytes_per_sec": 0, 00:14:54.981 "r_mbytes_per_sec": 0, 00:14:54.981 "w_mbytes_per_sec": 0 00:14:54.981 }, 00:14:54.981 "claimed": true, 00:14:54.981 "claim_type": "exclusive_write", 00:14:54.981 "zoned": false, 00:14:54.981 "supported_io_types": { 00:14:54.981 "read": true, 00:14:54.981 "write": true, 00:14:54.981 "unmap": true, 00:14:54.981 "flush": true, 00:14:54.981 "reset": true, 00:14:54.981 "nvme_admin": false, 00:14:54.981 "nvme_io": false, 00:14:54.981 "nvme_io_md": false, 00:14:54.981 "write_zeroes": true, 00:14:54.981 "zcopy": true, 00:14:54.981 "get_zone_info": false, 00:14:54.981 "zone_management": false, 00:14:54.981 "zone_append": false, 00:14:54.981 "compare": false, 00:14:54.981 "compare_and_write": false, 00:14:54.981 "abort": true, 00:14:54.981 "seek_hole": false, 00:14:54.981 "seek_data": false, 00:14:54.981 "copy": true, 00:14:54.981 "nvme_iov_md": false 00:14:54.981 }, 00:14:54.981 "memory_domains": [ 00:14:54.981 { 00:14:54.981 "dma_device_id": "system", 00:14:54.981 "dma_device_type": 1 00:14:54.981 }, 00:14:54.981 { 00:14:54.981 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:54.981 "dma_device_type": 2 00:14:54.981 } 00:14:54.981 ], 00:14:54.981 "driver_specific": {} 00:14:54.981 } 00:14:54.981 ] 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.981 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.982 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:54.982 "name": "Existed_Raid", 00:14:54.982 "uuid": "fd7d2bdb-760c-4a98-bd49-141cabe5d3da", 00:14:54.982 "strip_size_kb": 64, 00:14:54.982 "state": "online", 00:14:54.982 "raid_level": "raid5f", 00:14:54.982 "superblock": true, 00:14:54.982 "num_base_bdevs": 3, 00:14:54.982 "num_base_bdevs_discovered": 3, 00:14:54.982 "num_base_bdevs_operational": 3, 00:14:54.982 "base_bdevs_list": [ 00:14:54.982 { 00:14:54.982 "name": "BaseBdev1", 00:14:54.982 "uuid": "5626b12b-413c-41c9-841d-a4bb1cc0e22f", 00:14:54.982 "is_configured": true, 00:14:54.982 "data_offset": 2048, 00:14:54.982 "data_size": 63488 00:14:54.982 }, 00:14:54.982 { 00:14:54.982 "name": "BaseBdev2", 00:14:54.982 "uuid": "1f3813f4-0ca5-4421-babc-f56f1666f9a0", 00:14:54.982 "is_configured": true, 00:14:54.982 "data_offset": 2048, 00:14:54.982 "data_size": 63488 00:14:54.982 }, 00:14:54.982 { 00:14:54.982 "name": "BaseBdev3", 00:14:54.982 "uuid": "8389cd3e-0526-4d8a-8d83-0ac82bfb3ef9", 00:14:54.982 "is_configured": true, 00:14:54.982 "data_offset": 2048, 00:14:54.982 "data_size": 63488 00:14:54.982 } 00:14:54.982 ] 00:14:54.982 }' 00:14:54.982 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:54.982 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.243 [2024-11-20 16:02:53.354213] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:55.243 "name": "Existed_Raid", 00:14:55.243 "aliases": [ 00:14:55.243 "fd7d2bdb-760c-4a98-bd49-141cabe5d3da" 00:14:55.243 ], 00:14:55.243 "product_name": "Raid Volume", 00:14:55.243 "block_size": 512, 00:14:55.243 "num_blocks": 126976, 00:14:55.243 "uuid": "fd7d2bdb-760c-4a98-bd49-141cabe5d3da", 00:14:55.243 "assigned_rate_limits": { 00:14:55.243 "rw_ios_per_sec": 0, 00:14:55.243 "rw_mbytes_per_sec": 0, 00:14:55.243 "r_mbytes_per_sec": 0, 00:14:55.243 "w_mbytes_per_sec": 0 00:14:55.243 }, 00:14:55.243 "claimed": false, 00:14:55.243 "zoned": false, 00:14:55.243 "supported_io_types": { 00:14:55.243 "read": true, 00:14:55.243 "write": true, 00:14:55.243 "unmap": false, 00:14:55.243 "flush": false, 00:14:55.243 "reset": true, 00:14:55.243 "nvme_admin": false, 00:14:55.243 "nvme_io": false, 00:14:55.243 "nvme_io_md": false, 00:14:55.243 "write_zeroes": true, 00:14:55.243 "zcopy": false, 00:14:55.243 "get_zone_info": false, 00:14:55.243 "zone_management": false, 00:14:55.243 "zone_append": false, 00:14:55.243 "compare": false, 00:14:55.243 "compare_and_write": false, 00:14:55.243 "abort": false, 00:14:55.243 "seek_hole": false, 00:14:55.243 "seek_data": false, 00:14:55.243 "copy": false, 00:14:55.243 "nvme_iov_md": false 00:14:55.243 }, 00:14:55.243 "driver_specific": { 00:14:55.243 "raid": { 00:14:55.243 "uuid": "fd7d2bdb-760c-4a98-bd49-141cabe5d3da", 00:14:55.243 "strip_size_kb": 64, 00:14:55.243 "state": "online", 00:14:55.243 "raid_level": "raid5f", 00:14:55.243 "superblock": true, 00:14:55.243 "num_base_bdevs": 3, 00:14:55.243 "num_base_bdevs_discovered": 3, 00:14:55.243 "num_base_bdevs_operational": 3, 00:14:55.243 "base_bdevs_list": [ 00:14:55.243 { 00:14:55.243 "name": "BaseBdev1", 00:14:55.243 "uuid": "5626b12b-413c-41c9-841d-a4bb1cc0e22f", 00:14:55.243 "is_configured": true, 00:14:55.243 "data_offset": 2048, 00:14:55.243 "data_size": 63488 00:14:55.243 }, 00:14:55.243 { 00:14:55.243 "name": "BaseBdev2", 00:14:55.243 "uuid": "1f3813f4-0ca5-4421-babc-f56f1666f9a0", 00:14:55.243 "is_configured": true, 00:14:55.243 "data_offset": 2048, 00:14:55.243 "data_size": 63488 00:14:55.243 }, 00:14:55.243 { 00:14:55.243 "name": "BaseBdev3", 00:14:55.243 "uuid": "8389cd3e-0526-4d8a-8d83-0ac82bfb3ef9", 00:14:55.243 "is_configured": true, 00:14:55.243 "data_offset": 2048, 00:14:55.243 "data_size": 63488 00:14:55.243 } 00:14:55.243 ] 00:14:55.243 } 00:14:55.243 } 00:14:55.243 }' 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:55.243 BaseBdev2 00:14:55.243 BaseBdev3' 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.243 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.504 [2024-11-20 16:02:53.554066] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.504 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:55.504 "name": "Existed_Raid", 00:14:55.504 "uuid": "fd7d2bdb-760c-4a98-bd49-141cabe5d3da", 00:14:55.504 "strip_size_kb": 64, 00:14:55.504 "state": "online", 00:14:55.504 "raid_level": "raid5f", 00:14:55.504 "superblock": true, 00:14:55.504 "num_base_bdevs": 3, 00:14:55.504 "num_base_bdevs_discovered": 2, 00:14:55.504 "num_base_bdevs_operational": 2, 00:14:55.504 "base_bdevs_list": [ 00:14:55.504 { 00:14:55.504 "name": null, 00:14:55.504 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:55.504 "is_configured": false, 00:14:55.504 "data_offset": 0, 00:14:55.504 "data_size": 63488 00:14:55.504 }, 00:14:55.504 { 00:14:55.504 "name": "BaseBdev2", 00:14:55.504 "uuid": "1f3813f4-0ca5-4421-babc-f56f1666f9a0", 00:14:55.504 "is_configured": true, 00:14:55.504 "data_offset": 2048, 00:14:55.504 "data_size": 63488 00:14:55.504 }, 00:14:55.504 { 00:14:55.504 "name": "BaseBdev3", 00:14:55.504 "uuid": "8389cd3e-0526-4d8a-8d83-0ac82bfb3ef9", 00:14:55.504 "is_configured": true, 00:14:55.504 "data_offset": 2048, 00:14:55.504 "data_size": 63488 00:14:55.505 } 00:14:55.505 ] 00:14:55.505 }' 00:14:55.505 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:55.505 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.765 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:55.765 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:55.765 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:55.765 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.765 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:55.765 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.765 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.765 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:55.765 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:55.765 16:02:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:55.765 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.765 16:02:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.765 [2024-11-20 16:02:53.981238] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:55.765 [2024-11-20 16:02:53.981371] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:56.026 [2024-11-20 16:02:54.041273] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:56.026 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.026 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:56.026 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:56.026 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.026 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:56.026 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.026 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.026 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.026 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:56.026 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:56.026 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:56.026 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.026 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.026 [2024-11-20 16:02:54.081315] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:56.026 [2024-11-20 16:02:54.081357] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:14:56.026 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.026 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:56.026 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.027 BaseBdev2 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.027 [ 00:14:56.027 { 00:14:56.027 "name": "BaseBdev2", 00:14:56.027 "aliases": [ 00:14:56.027 "6f795113-d646-41fe-b526-a64e8c5a00f8" 00:14:56.027 ], 00:14:56.027 "product_name": "Malloc disk", 00:14:56.027 "block_size": 512, 00:14:56.027 "num_blocks": 65536, 00:14:56.027 "uuid": "6f795113-d646-41fe-b526-a64e8c5a00f8", 00:14:56.027 "assigned_rate_limits": { 00:14:56.027 "rw_ios_per_sec": 0, 00:14:56.027 "rw_mbytes_per_sec": 0, 00:14:56.027 "r_mbytes_per_sec": 0, 00:14:56.027 "w_mbytes_per_sec": 0 00:14:56.027 }, 00:14:56.027 "claimed": false, 00:14:56.027 "zoned": false, 00:14:56.027 "supported_io_types": { 00:14:56.027 "read": true, 00:14:56.027 "write": true, 00:14:56.027 "unmap": true, 00:14:56.027 "flush": true, 00:14:56.027 "reset": true, 00:14:56.027 "nvme_admin": false, 00:14:56.027 "nvme_io": false, 00:14:56.027 "nvme_io_md": false, 00:14:56.027 "write_zeroes": true, 00:14:56.027 "zcopy": true, 00:14:56.027 "get_zone_info": false, 00:14:56.027 "zone_management": false, 00:14:56.027 "zone_append": false, 00:14:56.027 "compare": false, 00:14:56.027 "compare_and_write": false, 00:14:56.027 "abort": true, 00:14:56.027 "seek_hole": false, 00:14:56.027 "seek_data": false, 00:14:56.027 "copy": true, 00:14:56.027 "nvme_iov_md": false 00:14:56.027 }, 00:14:56.027 "memory_domains": [ 00:14:56.027 { 00:14:56.027 "dma_device_id": "system", 00:14:56.027 "dma_device_type": 1 00:14:56.027 }, 00:14:56.027 { 00:14:56.027 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:56.027 "dma_device_type": 2 00:14:56.027 } 00:14:56.027 ], 00:14:56.027 "driver_specific": {} 00:14:56.027 } 00:14:56.027 ] 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.027 BaseBdev3 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.027 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.318 [ 00:14:56.318 { 00:14:56.318 "name": "BaseBdev3", 00:14:56.318 "aliases": [ 00:14:56.318 "21b65069-09ae-41ea-b845-8d491f6f1314" 00:14:56.318 ], 00:14:56.318 "product_name": "Malloc disk", 00:14:56.318 "block_size": 512, 00:14:56.318 "num_blocks": 65536, 00:14:56.318 "uuid": "21b65069-09ae-41ea-b845-8d491f6f1314", 00:14:56.318 "assigned_rate_limits": { 00:14:56.318 "rw_ios_per_sec": 0, 00:14:56.318 "rw_mbytes_per_sec": 0, 00:14:56.318 "r_mbytes_per_sec": 0, 00:14:56.318 "w_mbytes_per_sec": 0 00:14:56.318 }, 00:14:56.318 "claimed": false, 00:14:56.318 "zoned": false, 00:14:56.318 "supported_io_types": { 00:14:56.318 "read": true, 00:14:56.318 "write": true, 00:14:56.318 "unmap": true, 00:14:56.318 "flush": true, 00:14:56.318 "reset": true, 00:14:56.318 "nvme_admin": false, 00:14:56.318 "nvme_io": false, 00:14:56.318 "nvme_io_md": false, 00:14:56.318 "write_zeroes": true, 00:14:56.318 "zcopy": true, 00:14:56.318 "get_zone_info": false, 00:14:56.318 "zone_management": false, 00:14:56.318 "zone_append": false, 00:14:56.318 "compare": false, 00:14:56.318 "compare_and_write": false, 00:14:56.318 "abort": true, 00:14:56.318 "seek_hole": false, 00:14:56.318 "seek_data": false, 00:14:56.318 "copy": true, 00:14:56.318 "nvme_iov_md": false 00:14:56.318 }, 00:14:56.318 "memory_domains": [ 00:14:56.318 { 00:14:56.318 "dma_device_id": "system", 00:14:56.318 "dma_device_type": 1 00:14:56.318 }, 00:14:56.318 { 00:14:56.318 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:56.318 "dma_device_type": 2 00:14:56.318 } 00:14:56.318 ], 00:14:56.318 "driver_specific": {} 00:14:56.318 } 00:14:56.318 ] 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.318 [2024-11-20 16:02:54.293622] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:56.318 [2024-11-20 16:02:54.293780] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:56.318 [2024-11-20 16:02:54.293862] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:56.318 [2024-11-20 16:02:54.295775] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:56.318 "name": "Existed_Raid", 00:14:56.318 "uuid": "578e8ac4-1182-4ec6-a4a6-2be2cd1db1fd", 00:14:56.318 "strip_size_kb": 64, 00:14:56.318 "state": "configuring", 00:14:56.318 "raid_level": "raid5f", 00:14:56.318 "superblock": true, 00:14:56.318 "num_base_bdevs": 3, 00:14:56.318 "num_base_bdevs_discovered": 2, 00:14:56.318 "num_base_bdevs_operational": 3, 00:14:56.318 "base_bdevs_list": [ 00:14:56.318 { 00:14:56.318 "name": "BaseBdev1", 00:14:56.318 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.318 "is_configured": false, 00:14:56.318 "data_offset": 0, 00:14:56.318 "data_size": 0 00:14:56.318 }, 00:14:56.318 { 00:14:56.318 "name": "BaseBdev2", 00:14:56.318 "uuid": "6f795113-d646-41fe-b526-a64e8c5a00f8", 00:14:56.318 "is_configured": true, 00:14:56.318 "data_offset": 2048, 00:14:56.318 "data_size": 63488 00:14:56.318 }, 00:14:56.318 { 00:14:56.318 "name": "BaseBdev3", 00:14:56.318 "uuid": "21b65069-09ae-41ea-b845-8d491f6f1314", 00:14:56.318 "is_configured": true, 00:14:56.318 "data_offset": 2048, 00:14:56.318 "data_size": 63488 00:14:56.318 } 00:14:56.318 ] 00:14:56.318 }' 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:56.318 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.581 [2024-11-20 16:02:54.625710] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:56.581 "name": "Existed_Raid", 00:14:56.581 "uuid": "578e8ac4-1182-4ec6-a4a6-2be2cd1db1fd", 00:14:56.581 "strip_size_kb": 64, 00:14:56.581 "state": "configuring", 00:14:56.581 "raid_level": "raid5f", 00:14:56.581 "superblock": true, 00:14:56.581 "num_base_bdevs": 3, 00:14:56.581 "num_base_bdevs_discovered": 1, 00:14:56.581 "num_base_bdevs_operational": 3, 00:14:56.581 "base_bdevs_list": [ 00:14:56.581 { 00:14:56.581 "name": "BaseBdev1", 00:14:56.581 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.581 "is_configured": false, 00:14:56.581 "data_offset": 0, 00:14:56.581 "data_size": 0 00:14:56.581 }, 00:14:56.581 { 00:14:56.581 "name": null, 00:14:56.581 "uuid": "6f795113-d646-41fe-b526-a64e8c5a00f8", 00:14:56.581 "is_configured": false, 00:14:56.581 "data_offset": 0, 00:14:56.581 "data_size": 63488 00:14:56.581 }, 00:14:56.581 { 00:14:56.581 "name": "BaseBdev3", 00:14:56.581 "uuid": "21b65069-09ae-41ea-b845-8d491f6f1314", 00:14:56.581 "is_configured": true, 00:14:56.581 "data_offset": 2048, 00:14:56.581 "data_size": 63488 00:14:56.581 } 00:14:56.581 ] 00:14:56.581 }' 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:56.581 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.844 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.844 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:56.844 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.844 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.844 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.844 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:56.844 16:02:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:56.844 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.844 16:02:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.844 [2024-11-20 16:02:55.016202] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:56.844 BaseBdev1 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.844 [ 00:14:56.844 { 00:14:56.844 "name": "BaseBdev1", 00:14:56.844 "aliases": [ 00:14:56.844 "632bf8c6-cf42-46d3-8e00-b28085b8debb" 00:14:56.844 ], 00:14:56.844 "product_name": "Malloc disk", 00:14:56.844 "block_size": 512, 00:14:56.844 "num_blocks": 65536, 00:14:56.844 "uuid": "632bf8c6-cf42-46d3-8e00-b28085b8debb", 00:14:56.844 "assigned_rate_limits": { 00:14:56.844 "rw_ios_per_sec": 0, 00:14:56.844 "rw_mbytes_per_sec": 0, 00:14:56.844 "r_mbytes_per_sec": 0, 00:14:56.844 "w_mbytes_per_sec": 0 00:14:56.844 }, 00:14:56.844 "claimed": true, 00:14:56.844 "claim_type": "exclusive_write", 00:14:56.844 "zoned": false, 00:14:56.844 "supported_io_types": { 00:14:56.844 "read": true, 00:14:56.844 "write": true, 00:14:56.844 "unmap": true, 00:14:56.844 "flush": true, 00:14:56.844 "reset": true, 00:14:56.844 "nvme_admin": false, 00:14:56.844 "nvme_io": false, 00:14:56.844 "nvme_io_md": false, 00:14:56.844 "write_zeroes": true, 00:14:56.844 "zcopy": true, 00:14:56.844 "get_zone_info": false, 00:14:56.844 "zone_management": false, 00:14:56.844 "zone_append": false, 00:14:56.844 "compare": false, 00:14:56.844 "compare_and_write": false, 00:14:56.844 "abort": true, 00:14:56.844 "seek_hole": false, 00:14:56.844 "seek_data": false, 00:14:56.844 "copy": true, 00:14:56.844 "nvme_iov_md": false 00:14:56.844 }, 00:14:56.844 "memory_domains": [ 00:14:56.844 { 00:14:56.844 "dma_device_id": "system", 00:14:56.844 "dma_device_type": 1 00:14:56.844 }, 00:14:56.844 { 00:14:56.844 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:56.844 "dma_device_type": 2 00:14:56.844 } 00:14:56.844 ], 00:14:56.844 "driver_specific": {} 00:14:56.844 } 00:14:56.844 ] 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:56.844 "name": "Existed_Raid", 00:14:56.844 "uuid": "578e8ac4-1182-4ec6-a4a6-2be2cd1db1fd", 00:14:56.844 "strip_size_kb": 64, 00:14:56.844 "state": "configuring", 00:14:56.844 "raid_level": "raid5f", 00:14:56.844 "superblock": true, 00:14:56.844 "num_base_bdevs": 3, 00:14:56.844 "num_base_bdevs_discovered": 2, 00:14:56.844 "num_base_bdevs_operational": 3, 00:14:56.844 "base_bdevs_list": [ 00:14:56.844 { 00:14:56.844 "name": "BaseBdev1", 00:14:56.844 "uuid": "632bf8c6-cf42-46d3-8e00-b28085b8debb", 00:14:56.844 "is_configured": true, 00:14:56.844 "data_offset": 2048, 00:14:56.844 "data_size": 63488 00:14:56.844 }, 00:14:56.844 { 00:14:56.844 "name": null, 00:14:56.844 "uuid": "6f795113-d646-41fe-b526-a64e8c5a00f8", 00:14:56.844 "is_configured": false, 00:14:56.844 "data_offset": 0, 00:14:56.844 "data_size": 63488 00:14:56.844 }, 00:14:56.844 { 00:14:56.844 "name": "BaseBdev3", 00:14:56.844 "uuid": "21b65069-09ae-41ea-b845-8d491f6f1314", 00:14:56.844 "is_configured": true, 00:14:56.844 "data_offset": 2048, 00:14:56.844 "data_size": 63488 00:14:56.844 } 00:14:56.844 ] 00:14:56.844 }' 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:56.844 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.417 [2024-11-20 16:02:55.396360] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:57.417 "name": "Existed_Raid", 00:14:57.417 "uuid": "578e8ac4-1182-4ec6-a4a6-2be2cd1db1fd", 00:14:57.417 "strip_size_kb": 64, 00:14:57.417 "state": "configuring", 00:14:57.417 "raid_level": "raid5f", 00:14:57.417 "superblock": true, 00:14:57.417 "num_base_bdevs": 3, 00:14:57.417 "num_base_bdevs_discovered": 1, 00:14:57.417 "num_base_bdevs_operational": 3, 00:14:57.417 "base_bdevs_list": [ 00:14:57.417 { 00:14:57.417 "name": "BaseBdev1", 00:14:57.417 "uuid": "632bf8c6-cf42-46d3-8e00-b28085b8debb", 00:14:57.417 "is_configured": true, 00:14:57.417 "data_offset": 2048, 00:14:57.417 "data_size": 63488 00:14:57.417 }, 00:14:57.417 { 00:14:57.417 "name": null, 00:14:57.417 "uuid": "6f795113-d646-41fe-b526-a64e8c5a00f8", 00:14:57.417 "is_configured": false, 00:14:57.417 "data_offset": 0, 00:14:57.417 "data_size": 63488 00:14:57.417 }, 00:14:57.417 { 00:14:57.417 "name": null, 00:14:57.417 "uuid": "21b65069-09ae-41ea-b845-8d491f6f1314", 00:14:57.417 "is_configured": false, 00:14:57.417 "data_offset": 0, 00:14:57.417 "data_size": 63488 00:14:57.417 } 00:14:57.417 ] 00:14:57.417 }' 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:57.417 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.677 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.677 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.677 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.677 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.678 [2024-11-20 16:02:55.760464] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:57.678 "name": "Existed_Raid", 00:14:57.678 "uuid": "578e8ac4-1182-4ec6-a4a6-2be2cd1db1fd", 00:14:57.678 "strip_size_kb": 64, 00:14:57.678 "state": "configuring", 00:14:57.678 "raid_level": "raid5f", 00:14:57.678 "superblock": true, 00:14:57.678 "num_base_bdevs": 3, 00:14:57.678 "num_base_bdevs_discovered": 2, 00:14:57.678 "num_base_bdevs_operational": 3, 00:14:57.678 "base_bdevs_list": [ 00:14:57.678 { 00:14:57.678 "name": "BaseBdev1", 00:14:57.678 "uuid": "632bf8c6-cf42-46d3-8e00-b28085b8debb", 00:14:57.678 "is_configured": true, 00:14:57.678 "data_offset": 2048, 00:14:57.678 "data_size": 63488 00:14:57.678 }, 00:14:57.678 { 00:14:57.678 "name": null, 00:14:57.678 "uuid": "6f795113-d646-41fe-b526-a64e8c5a00f8", 00:14:57.678 "is_configured": false, 00:14:57.678 "data_offset": 0, 00:14:57.678 "data_size": 63488 00:14:57.678 }, 00:14:57.678 { 00:14:57.678 "name": "BaseBdev3", 00:14:57.678 "uuid": "21b65069-09ae-41ea-b845-8d491f6f1314", 00:14:57.678 "is_configured": true, 00:14:57.678 "data_offset": 2048, 00:14:57.678 "data_size": 63488 00:14:57.678 } 00:14:57.678 ] 00:14:57.678 }' 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:57.678 16:02:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.938 [2024-11-20 16:02:56.116556] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.938 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.200 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.200 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:58.200 "name": "Existed_Raid", 00:14:58.200 "uuid": "578e8ac4-1182-4ec6-a4a6-2be2cd1db1fd", 00:14:58.200 "strip_size_kb": 64, 00:14:58.200 "state": "configuring", 00:14:58.200 "raid_level": "raid5f", 00:14:58.200 "superblock": true, 00:14:58.200 "num_base_bdevs": 3, 00:14:58.200 "num_base_bdevs_discovered": 1, 00:14:58.200 "num_base_bdevs_operational": 3, 00:14:58.200 "base_bdevs_list": [ 00:14:58.200 { 00:14:58.200 "name": null, 00:14:58.200 "uuid": "632bf8c6-cf42-46d3-8e00-b28085b8debb", 00:14:58.200 "is_configured": false, 00:14:58.200 "data_offset": 0, 00:14:58.200 "data_size": 63488 00:14:58.200 }, 00:14:58.200 { 00:14:58.200 "name": null, 00:14:58.200 "uuid": "6f795113-d646-41fe-b526-a64e8c5a00f8", 00:14:58.200 "is_configured": false, 00:14:58.200 "data_offset": 0, 00:14:58.200 "data_size": 63488 00:14:58.200 }, 00:14:58.200 { 00:14:58.200 "name": "BaseBdev3", 00:14:58.200 "uuid": "21b65069-09ae-41ea-b845-8d491f6f1314", 00:14:58.200 "is_configured": true, 00:14:58.200 "data_offset": 2048, 00:14:58.200 "data_size": 63488 00:14:58.200 } 00:14:58.200 ] 00:14:58.200 }' 00:14:58.200 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:58.200 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.462 [2024-11-20 16:02:56.559390] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:58.462 "name": "Existed_Raid", 00:14:58.462 "uuid": "578e8ac4-1182-4ec6-a4a6-2be2cd1db1fd", 00:14:58.462 "strip_size_kb": 64, 00:14:58.462 "state": "configuring", 00:14:58.462 "raid_level": "raid5f", 00:14:58.462 "superblock": true, 00:14:58.462 "num_base_bdevs": 3, 00:14:58.462 "num_base_bdevs_discovered": 2, 00:14:58.462 "num_base_bdevs_operational": 3, 00:14:58.462 "base_bdevs_list": [ 00:14:58.462 { 00:14:58.462 "name": null, 00:14:58.462 "uuid": "632bf8c6-cf42-46d3-8e00-b28085b8debb", 00:14:58.462 "is_configured": false, 00:14:58.462 "data_offset": 0, 00:14:58.462 "data_size": 63488 00:14:58.462 }, 00:14:58.462 { 00:14:58.462 "name": "BaseBdev2", 00:14:58.462 "uuid": "6f795113-d646-41fe-b526-a64e8c5a00f8", 00:14:58.462 "is_configured": true, 00:14:58.462 "data_offset": 2048, 00:14:58.462 "data_size": 63488 00:14:58.462 }, 00:14:58.462 { 00:14:58.462 "name": "BaseBdev3", 00:14:58.462 "uuid": "21b65069-09ae-41ea-b845-8d491f6f1314", 00:14:58.462 "is_configured": true, 00:14:58.462 "data_offset": 2048, 00:14:58.462 "data_size": 63488 00:14:58.462 } 00:14:58.462 ] 00:14:58.462 }' 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:58.462 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.724 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.724 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.724 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.724 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:58.724 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.724 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:58.724 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:58.724 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.724 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.725 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.725 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.725 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 632bf8c6-cf42-46d3-8e00-b28085b8debb 00:14:58.725 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.725 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.725 [2024-11-20 16:02:56.966009] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:58.725 NewBaseBdev 00:14:58.725 [2024-11-20 16:02:56.966324] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:14:58.725 [2024-11-20 16:02:56.966345] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:58.725 [2024-11-20 16:02:56.966591] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:14:58.725 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.725 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:58.725 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:14:58.725 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:58.725 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:58.725 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:58.725 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:58.725 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:58.725 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.725 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.725 [2024-11-20 16:02:56.970095] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:14:58.725 [2024-11-20 16:02:56.970113] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:14:58.725 [2024-11-20 16:02:56.970248] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:58.986 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.986 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:58.986 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.986 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.986 [ 00:14:58.986 { 00:14:58.986 "name": "NewBaseBdev", 00:14:58.986 "aliases": [ 00:14:58.986 "632bf8c6-cf42-46d3-8e00-b28085b8debb" 00:14:58.986 ], 00:14:58.986 "product_name": "Malloc disk", 00:14:58.986 "block_size": 512, 00:14:58.986 "num_blocks": 65536, 00:14:58.986 "uuid": "632bf8c6-cf42-46d3-8e00-b28085b8debb", 00:14:58.986 "assigned_rate_limits": { 00:14:58.986 "rw_ios_per_sec": 0, 00:14:58.986 "rw_mbytes_per_sec": 0, 00:14:58.986 "r_mbytes_per_sec": 0, 00:14:58.986 "w_mbytes_per_sec": 0 00:14:58.986 }, 00:14:58.986 "claimed": true, 00:14:58.986 "claim_type": "exclusive_write", 00:14:58.986 "zoned": false, 00:14:58.986 "supported_io_types": { 00:14:58.986 "read": true, 00:14:58.986 "write": true, 00:14:58.986 "unmap": true, 00:14:58.986 "flush": true, 00:14:58.986 "reset": true, 00:14:58.986 "nvme_admin": false, 00:14:58.986 "nvme_io": false, 00:14:58.986 "nvme_io_md": false, 00:14:58.986 "write_zeroes": true, 00:14:58.986 "zcopy": true, 00:14:58.986 "get_zone_info": false, 00:14:58.986 "zone_management": false, 00:14:58.986 "zone_append": false, 00:14:58.987 "compare": false, 00:14:58.987 "compare_and_write": false, 00:14:58.987 "abort": true, 00:14:58.987 "seek_hole": false, 00:14:58.987 "seek_data": false, 00:14:58.987 "copy": true, 00:14:58.987 "nvme_iov_md": false 00:14:58.987 }, 00:14:58.987 "memory_domains": [ 00:14:58.987 { 00:14:58.987 "dma_device_id": "system", 00:14:58.987 "dma_device_type": 1 00:14:58.987 }, 00:14:58.987 { 00:14:58.987 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:58.987 "dma_device_type": 2 00:14:58.987 } 00:14:58.987 ], 00:14:58.987 "driver_specific": {} 00:14:58.987 } 00:14:58.987 ] 00:14:58.987 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.987 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:58.987 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:58.987 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:58.987 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:58.987 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:58.987 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:58.987 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:58.987 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:58.987 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:58.987 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:58.987 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:58.987 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.987 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.987 16:02:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.987 16:02:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:58.987 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.987 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:58.987 "name": "Existed_Raid", 00:14:58.987 "uuid": "578e8ac4-1182-4ec6-a4a6-2be2cd1db1fd", 00:14:58.987 "strip_size_kb": 64, 00:14:58.987 "state": "online", 00:14:58.987 "raid_level": "raid5f", 00:14:58.987 "superblock": true, 00:14:58.987 "num_base_bdevs": 3, 00:14:58.987 "num_base_bdevs_discovered": 3, 00:14:58.987 "num_base_bdevs_operational": 3, 00:14:58.987 "base_bdevs_list": [ 00:14:58.987 { 00:14:58.987 "name": "NewBaseBdev", 00:14:58.987 "uuid": "632bf8c6-cf42-46d3-8e00-b28085b8debb", 00:14:58.987 "is_configured": true, 00:14:58.987 "data_offset": 2048, 00:14:58.987 "data_size": 63488 00:14:58.987 }, 00:14:58.987 { 00:14:58.987 "name": "BaseBdev2", 00:14:58.987 "uuid": "6f795113-d646-41fe-b526-a64e8c5a00f8", 00:14:58.987 "is_configured": true, 00:14:58.987 "data_offset": 2048, 00:14:58.987 "data_size": 63488 00:14:58.987 }, 00:14:58.987 { 00:14:58.987 "name": "BaseBdev3", 00:14:58.987 "uuid": "21b65069-09ae-41ea-b845-8d491f6f1314", 00:14:58.987 "is_configured": true, 00:14:58.987 "data_offset": 2048, 00:14:58.987 "data_size": 63488 00:14:58.987 } 00:14:58.987 ] 00:14:58.987 }' 00:14:58.987 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:58.987 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.249 [2024-11-20 16:02:57.318518] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:59.249 "name": "Existed_Raid", 00:14:59.249 "aliases": [ 00:14:59.249 "578e8ac4-1182-4ec6-a4a6-2be2cd1db1fd" 00:14:59.249 ], 00:14:59.249 "product_name": "Raid Volume", 00:14:59.249 "block_size": 512, 00:14:59.249 "num_blocks": 126976, 00:14:59.249 "uuid": "578e8ac4-1182-4ec6-a4a6-2be2cd1db1fd", 00:14:59.249 "assigned_rate_limits": { 00:14:59.249 "rw_ios_per_sec": 0, 00:14:59.249 "rw_mbytes_per_sec": 0, 00:14:59.249 "r_mbytes_per_sec": 0, 00:14:59.249 "w_mbytes_per_sec": 0 00:14:59.249 }, 00:14:59.249 "claimed": false, 00:14:59.249 "zoned": false, 00:14:59.249 "supported_io_types": { 00:14:59.249 "read": true, 00:14:59.249 "write": true, 00:14:59.249 "unmap": false, 00:14:59.249 "flush": false, 00:14:59.249 "reset": true, 00:14:59.249 "nvme_admin": false, 00:14:59.249 "nvme_io": false, 00:14:59.249 "nvme_io_md": false, 00:14:59.249 "write_zeroes": true, 00:14:59.249 "zcopy": false, 00:14:59.249 "get_zone_info": false, 00:14:59.249 "zone_management": false, 00:14:59.249 "zone_append": false, 00:14:59.249 "compare": false, 00:14:59.249 "compare_and_write": false, 00:14:59.249 "abort": false, 00:14:59.249 "seek_hole": false, 00:14:59.249 "seek_data": false, 00:14:59.249 "copy": false, 00:14:59.249 "nvme_iov_md": false 00:14:59.249 }, 00:14:59.249 "driver_specific": { 00:14:59.249 "raid": { 00:14:59.249 "uuid": "578e8ac4-1182-4ec6-a4a6-2be2cd1db1fd", 00:14:59.249 "strip_size_kb": 64, 00:14:59.249 "state": "online", 00:14:59.249 "raid_level": "raid5f", 00:14:59.249 "superblock": true, 00:14:59.249 "num_base_bdevs": 3, 00:14:59.249 "num_base_bdevs_discovered": 3, 00:14:59.249 "num_base_bdevs_operational": 3, 00:14:59.249 "base_bdevs_list": [ 00:14:59.249 { 00:14:59.249 "name": "NewBaseBdev", 00:14:59.249 "uuid": "632bf8c6-cf42-46d3-8e00-b28085b8debb", 00:14:59.249 "is_configured": true, 00:14:59.249 "data_offset": 2048, 00:14:59.249 "data_size": 63488 00:14:59.249 }, 00:14:59.249 { 00:14:59.249 "name": "BaseBdev2", 00:14:59.249 "uuid": "6f795113-d646-41fe-b526-a64e8c5a00f8", 00:14:59.249 "is_configured": true, 00:14:59.249 "data_offset": 2048, 00:14:59.249 "data_size": 63488 00:14:59.249 }, 00:14:59.249 { 00:14:59.249 "name": "BaseBdev3", 00:14:59.249 "uuid": "21b65069-09ae-41ea-b845-8d491f6f1314", 00:14:59.249 "is_configured": true, 00:14:59.249 "data_offset": 2048, 00:14:59.249 "data_size": 63488 00:14:59.249 } 00:14:59.249 ] 00:14:59.249 } 00:14:59.249 } 00:14:59.249 }' 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:59.249 BaseBdev2 00:14:59.249 BaseBdev3' 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.249 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.510 [2024-11-20 16:02:57.534363] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:59.510 [2024-11-20 16:02:57.534387] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:59.510 [2024-11-20 16:02:57.534459] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:59.510 [2024-11-20 16:02:57.534756] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:59.510 [2024-11-20 16:02:57.534770] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 78316 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 78316 ']' 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 78316 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 78316 00:14:59.510 killing process with pid 78316 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 78316' 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 78316 00:14:59.510 [2024-11-20 16:02:57.566599] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:59.510 16:02:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 78316 00:14:59.510 [2024-11-20 16:02:57.754014] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:00.451 16:02:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:15:00.451 00:15:00.451 real 0m7.894s 00:15:00.451 user 0m12.515s 00:15:00.451 sys 0m1.295s 00:15:00.451 16:02:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:00.451 ************************************ 00:15:00.451 END TEST raid5f_state_function_test_sb 00:15:00.451 ************************************ 00:15:00.451 16:02:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.451 16:02:58 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 3 00:15:00.451 16:02:58 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:15:00.451 16:02:58 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:00.451 16:02:58 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:00.451 ************************************ 00:15:00.451 START TEST raid5f_superblock_test 00:15:00.451 ************************************ 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid5f 3 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=78913 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 78913 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 78913 ']' 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:00.451 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.451 16:02:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:15:00.451 [2024-11-20 16:02:58.619811] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:15:00.451 [2024-11-20 16:02:58.619935] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid78913 ] 00:15:00.711 [2024-11-20 16:02:58.772389] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:00.711 [2024-11-20 16:02:58.875586] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:00.970 [2024-11-20 16:02:59.013203] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:00.970 [2024-11-20 16:02:59.013245] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:01.230 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:01.230 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:15:01.230 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:15:01.230 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:01.230 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:15:01.230 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:15:01.230 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:15:01.230 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:01.230 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:01.230 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:01.230 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:15:01.230 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:01.230 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.490 malloc1 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.490 [2024-11-20 16:02:59.497113] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:01.490 [2024-11-20 16:02:59.497174] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:01.490 [2024-11-20 16:02:59.497195] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:01.490 [2024-11-20 16:02:59.497205] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:01.490 [2024-11-20 16:02:59.499399] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:01.490 [2024-11-20 16:02:59.499562] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:01.490 pt1 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.490 malloc2 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.490 [2024-11-20 16:02:59.537576] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:01.490 [2024-11-20 16:02:59.537631] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:01.490 [2024-11-20 16:02:59.537655] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:01.490 [2024-11-20 16:02:59.537664] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:01.490 [2024-11-20 16:02:59.539867] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:01.490 [2024-11-20 16:02:59.539909] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:01.490 pt2 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:01.490 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.490 malloc3 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.491 [2024-11-20 16:02:59.591178] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:15:01.491 [2024-11-20 16:02:59.591230] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:01.491 [2024-11-20 16:02:59.591255] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:15:01.491 [2024-11-20 16:02:59.591265] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:01.491 [2024-11-20 16:02:59.593366] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:01.491 [2024-11-20 16:02:59.593503] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:15:01.491 pt3 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.491 [2024-11-20 16:02:59.599225] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:01.491 [2024-11-20 16:02:59.601179] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:01.491 [2024-11-20 16:02:59.601323] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:01.491 [2024-11-20 16:02:59.601512] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:15:01.491 [2024-11-20 16:02:59.601603] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:15:01.491 [2024-11-20 16:02:59.601899] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:15:01.491 [2024-11-20 16:02:59.605652] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:15:01.491 [2024-11-20 16:02:59.605757] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:15:01.491 [2024-11-20 16:02:59.605997] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:01.491 "name": "raid_bdev1", 00:15:01.491 "uuid": "c8de4767-25b8-486a-8a7f-12aa6ee9cf59", 00:15:01.491 "strip_size_kb": 64, 00:15:01.491 "state": "online", 00:15:01.491 "raid_level": "raid5f", 00:15:01.491 "superblock": true, 00:15:01.491 "num_base_bdevs": 3, 00:15:01.491 "num_base_bdevs_discovered": 3, 00:15:01.491 "num_base_bdevs_operational": 3, 00:15:01.491 "base_bdevs_list": [ 00:15:01.491 { 00:15:01.491 "name": "pt1", 00:15:01.491 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:01.491 "is_configured": true, 00:15:01.491 "data_offset": 2048, 00:15:01.491 "data_size": 63488 00:15:01.491 }, 00:15:01.491 { 00:15:01.491 "name": "pt2", 00:15:01.491 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:01.491 "is_configured": true, 00:15:01.491 "data_offset": 2048, 00:15:01.491 "data_size": 63488 00:15:01.491 }, 00:15:01.491 { 00:15:01.491 "name": "pt3", 00:15:01.491 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:01.491 "is_configured": true, 00:15:01.491 "data_offset": 2048, 00:15:01.491 "data_size": 63488 00:15:01.491 } 00:15:01.491 ] 00:15:01.491 }' 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:01.491 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.751 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:15:01.751 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:01.751 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:01.751 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:01.751 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:15:01.751 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:01.752 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:01.752 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:01.752 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.752 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:01.752 [2024-11-20 16:02:59.938387] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:01.752 16:02:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:01.752 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:01.752 "name": "raid_bdev1", 00:15:01.752 "aliases": [ 00:15:01.752 "c8de4767-25b8-486a-8a7f-12aa6ee9cf59" 00:15:01.752 ], 00:15:01.752 "product_name": "Raid Volume", 00:15:01.752 "block_size": 512, 00:15:01.752 "num_blocks": 126976, 00:15:01.752 "uuid": "c8de4767-25b8-486a-8a7f-12aa6ee9cf59", 00:15:01.752 "assigned_rate_limits": { 00:15:01.752 "rw_ios_per_sec": 0, 00:15:01.752 "rw_mbytes_per_sec": 0, 00:15:01.752 "r_mbytes_per_sec": 0, 00:15:01.752 "w_mbytes_per_sec": 0 00:15:01.752 }, 00:15:01.752 "claimed": false, 00:15:01.752 "zoned": false, 00:15:01.752 "supported_io_types": { 00:15:01.752 "read": true, 00:15:01.752 "write": true, 00:15:01.752 "unmap": false, 00:15:01.752 "flush": false, 00:15:01.752 "reset": true, 00:15:01.752 "nvme_admin": false, 00:15:01.752 "nvme_io": false, 00:15:01.752 "nvme_io_md": false, 00:15:01.752 "write_zeroes": true, 00:15:01.752 "zcopy": false, 00:15:01.752 "get_zone_info": false, 00:15:01.752 "zone_management": false, 00:15:01.752 "zone_append": false, 00:15:01.752 "compare": false, 00:15:01.752 "compare_and_write": false, 00:15:01.752 "abort": false, 00:15:01.752 "seek_hole": false, 00:15:01.752 "seek_data": false, 00:15:01.752 "copy": false, 00:15:01.752 "nvme_iov_md": false 00:15:01.752 }, 00:15:01.752 "driver_specific": { 00:15:01.752 "raid": { 00:15:01.752 "uuid": "c8de4767-25b8-486a-8a7f-12aa6ee9cf59", 00:15:01.752 "strip_size_kb": 64, 00:15:01.752 "state": "online", 00:15:01.752 "raid_level": "raid5f", 00:15:01.752 "superblock": true, 00:15:01.752 "num_base_bdevs": 3, 00:15:01.752 "num_base_bdevs_discovered": 3, 00:15:01.752 "num_base_bdevs_operational": 3, 00:15:01.752 "base_bdevs_list": [ 00:15:01.752 { 00:15:01.752 "name": "pt1", 00:15:01.752 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:01.752 "is_configured": true, 00:15:01.752 "data_offset": 2048, 00:15:01.752 "data_size": 63488 00:15:01.752 }, 00:15:01.752 { 00:15:01.752 "name": "pt2", 00:15:01.752 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:01.752 "is_configured": true, 00:15:01.752 "data_offset": 2048, 00:15:01.752 "data_size": 63488 00:15:01.752 }, 00:15:01.752 { 00:15:01.752 "name": "pt3", 00:15:01.752 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:01.752 "is_configured": true, 00:15:01.752 "data_offset": 2048, 00:15:01.752 "data_size": 63488 00:15:01.752 } 00:15:01.752 ] 00:15:01.752 } 00:15:01.752 } 00:15:01.752 }' 00:15:01.752 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:01.752 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:01.752 pt2 00:15:01.752 pt3' 00:15:01.752 16:02:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:02.013 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:02.013 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:02.013 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:02.013 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.013 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.013 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:02.013 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.013 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:02.013 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:02.013 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:02.013 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:02.013 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:02.013 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.014 [2024-11-20 16:03:00.138392] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=c8de4767-25b8-486a-8a7f-12aa6ee9cf59 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z c8de4767-25b8-486a-8a7f-12aa6ee9cf59 ']' 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.014 [2024-11-20 16:03:00.162192] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:02.014 [2024-11-20 16:03:00.162214] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:02.014 [2024-11-20 16:03:00.162278] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:02.014 [2024-11-20 16:03:00.162354] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:02.014 [2024-11-20 16:03:00.162365] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:15:02.014 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.342 [2024-11-20 16:03:00.266296] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:15:02.342 [2024-11-20 16:03:00.268244] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:15:02.342 [2024-11-20 16:03:00.268407] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:15:02.342 [2024-11-20 16:03:00.268467] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:15:02.342 [2024-11-20 16:03:00.268516] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:15:02.342 [2024-11-20 16:03:00.268537] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:15:02.342 [2024-11-20 16:03:00.268554] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:02.342 [2024-11-20 16:03:00.268565] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:15:02.342 request: 00:15:02.342 { 00:15:02.342 "name": "raid_bdev1", 00:15:02.342 "raid_level": "raid5f", 00:15:02.342 "base_bdevs": [ 00:15:02.342 "malloc1", 00:15:02.342 "malloc2", 00:15:02.342 "malloc3" 00:15:02.342 ], 00:15:02.342 "strip_size_kb": 64, 00:15:02.342 "superblock": false, 00:15:02.342 "method": "bdev_raid_create", 00:15:02.342 "req_id": 1 00:15:02.342 } 00:15:02.342 Got JSON-RPC error response 00:15:02.342 response: 00:15:02.342 { 00:15:02.342 "code": -17, 00:15:02.342 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:15:02.342 } 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.342 [2024-11-20 16:03:00.310253] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:02.342 [2024-11-20 16:03:00.310305] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:02.342 [2024-11-20 16:03:00.310324] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:15:02.342 [2024-11-20 16:03:00.310332] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:02.342 [2024-11-20 16:03:00.312555] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:02.342 [2024-11-20 16:03:00.312590] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:02.342 [2024-11-20 16:03:00.312685] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:02.342 [2024-11-20 16:03:00.312734] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:02.342 pt1 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:02.342 "name": "raid_bdev1", 00:15:02.342 "uuid": "c8de4767-25b8-486a-8a7f-12aa6ee9cf59", 00:15:02.342 "strip_size_kb": 64, 00:15:02.342 "state": "configuring", 00:15:02.342 "raid_level": "raid5f", 00:15:02.342 "superblock": true, 00:15:02.342 "num_base_bdevs": 3, 00:15:02.342 "num_base_bdevs_discovered": 1, 00:15:02.342 "num_base_bdevs_operational": 3, 00:15:02.342 "base_bdevs_list": [ 00:15:02.342 { 00:15:02.342 "name": "pt1", 00:15:02.342 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:02.342 "is_configured": true, 00:15:02.342 "data_offset": 2048, 00:15:02.342 "data_size": 63488 00:15:02.342 }, 00:15:02.342 { 00:15:02.342 "name": null, 00:15:02.342 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:02.342 "is_configured": false, 00:15:02.342 "data_offset": 2048, 00:15:02.342 "data_size": 63488 00:15:02.342 }, 00:15:02.342 { 00:15:02.342 "name": null, 00:15:02.342 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:02.342 "is_configured": false, 00:15:02.342 "data_offset": 2048, 00:15:02.342 "data_size": 63488 00:15:02.342 } 00:15:02.342 ] 00:15:02.342 }' 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:02.342 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.617 [2024-11-20 16:03:00.630338] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:02.617 [2024-11-20 16:03:00.630504] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:02.617 [2024-11-20 16:03:00.630533] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:15:02.617 [2024-11-20 16:03:00.630543] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:02.617 [2024-11-20 16:03:00.630951] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:02.617 [2024-11-20 16:03:00.630971] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:02.617 [2024-11-20 16:03:00.631048] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:02.617 [2024-11-20 16:03:00.631072] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:02.617 pt2 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.617 [2024-11-20 16:03:00.638342] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:02.617 "name": "raid_bdev1", 00:15:02.617 "uuid": "c8de4767-25b8-486a-8a7f-12aa6ee9cf59", 00:15:02.617 "strip_size_kb": 64, 00:15:02.617 "state": "configuring", 00:15:02.617 "raid_level": "raid5f", 00:15:02.617 "superblock": true, 00:15:02.617 "num_base_bdevs": 3, 00:15:02.617 "num_base_bdevs_discovered": 1, 00:15:02.617 "num_base_bdevs_operational": 3, 00:15:02.617 "base_bdevs_list": [ 00:15:02.617 { 00:15:02.617 "name": "pt1", 00:15:02.617 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:02.617 "is_configured": true, 00:15:02.617 "data_offset": 2048, 00:15:02.617 "data_size": 63488 00:15:02.617 }, 00:15:02.617 { 00:15:02.617 "name": null, 00:15:02.617 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:02.617 "is_configured": false, 00:15:02.617 "data_offset": 0, 00:15:02.617 "data_size": 63488 00:15:02.617 }, 00:15:02.617 { 00:15:02.617 "name": null, 00:15:02.617 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:02.617 "is_configured": false, 00:15:02.617 "data_offset": 2048, 00:15:02.617 "data_size": 63488 00:15:02.617 } 00:15:02.617 ] 00:15:02.617 }' 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:02.617 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.879 [2024-11-20 16:03:00.970400] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:02.879 [2024-11-20 16:03:00.970463] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:02.879 [2024-11-20 16:03:00.970479] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:15:02.879 [2024-11-20 16:03:00.970490] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:02.879 [2024-11-20 16:03:00.970919] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:02.879 [2024-11-20 16:03:00.970942] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:02.879 [2024-11-20 16:03:00.971012] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:02.879 [2024-11-20 16:03:00.971039] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:02.879 pt2 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.879 [2024-11-20 16:03:00.978392] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:15:02.879 [2024-11-20 16:03:00.978518] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:02.879 [2024-11-20 16:03:00.978554] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:15:02.879 [2024-11-20 16:03:00.978616] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:02.879 [2024-11-20 16:03:00.979004] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:02.879 [2024-11-20 16:03:00.979213] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:15:02.879 [2024-11-20 16:03:00.979304] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:15:02.879 [2024-11-20 16:03:00.979345] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:02.879 [2024-11-20 16:03:00.979482] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:15:02.879 [2024-11-20 16:03:00.979512] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:15:02.879 [2024-11-20 16:03:00.979776] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:15:02.879 [2024-11-20 16:03:00.983312] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:15:02.879 [2024-11-20 16:03:00.983400] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:15:02.879 [2024-11-20 16:03:00.983615] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:02.879 pt3 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.879 16:03:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.879 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.879 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:02.879 "name": "raid_bdev1", 00:15:02.879 "uuid": "c8de4767-25b8-486a-8a7f-12aa6ee9cf59", 00:15:02.879 "strip_size_kb": 64, 00:15:02.879 "state": "online", 00:15:02.879 "raid_level": "raid5f", 00:15:02.879 "superblock": true, 00:15:02.879 "num_base_bdevs": 3, 00:15:02.879 "num_base_bdevs_discovered": 3, 00:15:02.879 "num_base_bdevs_operational": 3, 00:15:02.879 "base_bdevs_list": [ 00:15:02.879 { 00:15:02.879 "name": "pt1", 00:15:02.879 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:02.879 "is_configured": true, 00:15:02.879 "data_offset": 2048, 00:15:02.879 "data_size": 63488 00:15:02.879 }, 00:15:02.879 { 00:15:02.879 "name": "pt2", 00:15:02.879 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:02.879 "is_configured": true, 00:15:02.879 "data_offset": 2048, 00:15:02.879 "data_size": 63488 00:15:02.879 }, 00:15:02.879 { 00:15:02.879 "name": "pt3", 00:15:02.879 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:02.879 "is_configured": true, 00:15:02.879 "data_offset": 2048, 00:15:02.879 "data_size": 63488 00:15:02.879 } 00:15:02.879 ] 00:15:02.879 }' 00:15:02.879 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:02.879 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.140 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:15:03.140 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:03.140 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:03.140 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:03.140 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:15:03.140 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:03.140 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:03.140 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:03.140 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.140 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.140 [2024-11-20 16:03:01.307930] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:03.140 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.140 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:03.140 "name": "raid_bdev1", 00:15:03.140 "aliases": [ 00:15:03.140 "c8de4767-25b8-486a-8a7f-12aa6ee9cf59" 00:15:03.140 ], 00:15:03.140 "product_name": "Raid Volume", 00:15:03.140 "block_size": 512, 00:15:03.140 "num_blocks": 126976, 00:15:03.140 "uuid": "c8de4767-25b8-486a-8a7f-12aa6ee9cf59", 00:15:03.140 "assigned_rate_limits": { 00:15:03.140 "rw_ios_per_sec": 0, 00:15:03.140 "rw_mbytes_per_sec": 0, 00:15:03.140 "r_mbytes_per_sec": 0, 00:15:03.140 "w_mbytes_per_sec": 0 00:15:03.140 }, 00:15:03.140 "claimed": false, 00:15:03.140 "zoned": false, 00:15:03.140 "supported_io_types": { 00:15:03.140 "read": true, 00:15:03.140 "write": true, 00:15:03.140 "unmap": false, 00:15:03.140 "flush": false, 00:15:03.140 "reset": true, 00:15:03.140 "nvme_admin": false, 00:15:03.140 "nvme_io": false, 00:15:03.140 "nvme_io_md": false, 00:15:03.140 "write_zeroes": true, 00:15:03.140 "zcopy": false, 00:15:03.140 "get_zone_info": false, 00:15:03.140 "zone_management": false, 00:15:03.140 "zone_append": false, 00:15:03.140 "compare": false, 00:15:03.140 "compare_and_write": false, 00:15:03.140 "abort": false, 00:15:03.140 "seek_hole": false, 00:15:03.140 "seek_data": false, 00:15:03.140 "copy": false, 00:15:03.140 "nvme_iov_md": false 00:15:03.140 }, 00:15:03.140 "driver_specific": { 00:15:03.140 "raid": { 00:15:03.140 "uuid": "c8de4767-25b8-486a-8a7f-12aa6ee9cf59", 00:15:03.140 "strip_size_kb": 64, 00:15:03.140 "state": "online", 00:15:03.140 "raid_level": "raid5f", 00:15:03.140 "superblock": true, 00:15:03.140 "num_base_bdevs": 3, 00:15:03.140 "num_base_bdevs_discovered": 3, 00:15:03.140 "num_base_bdevs_operational": 3, 00:15:03.140 "base_bdevs_list": [ 00:15:03.140 { 00:15:03.140 "name": "pt1", 00:15:03.140 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:03.140 "is_configured": true, 00:15:03.140 "data_offset": 2048, 00:15:03.140 "data_size": 63488 00:15:03.140 }, 00:15:03.140 { 00:15:03.140 "name": "pt2", 00:15:03.140 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:03.140 "is_configured": true, 00:15:03.140 "data_offset": 2048, 00:15:03.140 "data_size": 63488 00:15:03.140 }, 00:15:03.140 { 00:15:03.140 "name": "pt3", 00:15:03.140 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:03.140 "is_configured": true, 00:15:03.140 "data_offset": 2048, 00:15:03.140 "data_size": 63488 00:15:03.140 } 00:15:03.140 ] 00:15:03.140 } 00:15:03.140 } 00:15:03.140 }' 00:15:03.140 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:03.140 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:03.140 pt2 00:15:03.140 pt3' 00:15:03.140 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:15:03.400 [2024-11-20 16:03:01.495917] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' c8de4767-25b8-486a-8a7f-12aa6ee9cf59 '!=' c8de4767-25b8-486a-8a7f-12aa6ee9cf59 ']' 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.400 [2024-11-20 16:03:01.527771] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.400 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:03.400 "name": "raid_bdev1", 00:15:03.400 "uuid": "c8de4767-25b8-486a-8a7f-12aa6ee9cf59", 00:15:03.400 "strip_size_kb": 64, 00:15:03.400 "state": "online", 00:15:03.400 "raid_level": "raid5f", 00:15:03.400 "superblock": true, 00:15:03.400 "num_base_bdevs": 3, 00:15:03.400 "num_base_bdevs_discovered": 2, 00:15:03.400 "num_base_bdevs_operational": 2, 00:15:03.400 "base_bdevs_list": [ 00:15:03.400 { 00:15:03.400 "name": null, 00:15:03.400 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:03.400 "is_configured": false, 00:15:03.401 "data_offset": 0, 00:15:03.401 "data_size": 63488 00:15:03.401 }, 00:15:03.401 { 00:15:03.401 "name": "pt2", 00:15:03.401 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:03.401 "is_configured": true, 00:15:03.401 "data_offset": 2048, 00:15:03.401 "data_size": 63488 00:15:03.401 }, 00:15:03.401 { 00:15:03.401 "name": "pt3", 00:15:03.401 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:03.401 "is_configured": true, 00:15:03.401 "data_offset": 2048, 00:15:03.401 "data_size": 63488 00:15:03.401 } 00:15:03.401 ] 00:15:03.401 }' 00:15:03.401 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:03.401 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.663 [2024-11-20 16:03:01.831807] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:03.663 [2024-11-20 16:03:01.831830] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:03.663 [2024-11-20 16:03:01.831889] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:03.663 [2024-11-20 16:03:01.831942] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:03.663 [2024-11-20 16:03:01.831955] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.663 [2024-11-20 16:03:01.895808] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:03.663 [2024-11-20 16:03:01.895858] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:03.663 [2024-11-20 16:03:01.895874] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:15:03.663 [2024-11-20 16:03:01.895885] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:03.663 [2024-11-20 16:03:01.898072] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:03.663 [2024-11-20 16:03:01.898107] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:03.663 [2024-11-20 16:03:01.898177] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:03.663 [2024-11-20 16:03:01.898220] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:03.663 pt2 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:03.663 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:03.664 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:03.664 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:03.664 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:03.664 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.664 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.664 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.664 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:03.924 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.924 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:03.924 "name": "raid_bdev1", 00:15:03.924 "uuid": "c8de4767-25b8-486a-8a7f-12aa6ee9cf59", 00:15:03.924 "strip_size_kb": 64, 00:15:03.924 "state": "configuring", 00:15:03.924 "raid_level": "raid5f", 00:15:03.924 "superblock": true, 00:15:03.924 "num_base_bdevs": 3, 00:15:03.924 "num_base_bdevs_discovered": 1, 00:15:03.924 "num_base_bdevs_operational": 2, 00:15:03.924 "base_bdevs_list": [ 00:15:03.924 { 00:15:03.924 "name": null, 00:15:03.924 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:03.924 "is_configured": false, 00:15:03.924 "data_offset": 2048, 00:15:03.924 "data_size": 63488 00:15:03.924 }, 00:15:03.924 { 00:15:03.924 "name": "pt2", 00:15:03.924 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:03.924 "is_configured": true, 00:15:03.924 "data_offset": 2048, 00:15:03.924 "data_size": 63488 00:15:03.924 }, 00:15:03.924 { 00:15:03.924 "name": null, 00:15:03.924 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:03.924 "is_configured": false, 00:15:03.924 "data_offset": 2048, 00:15:03.924 "data_size": 63488 00:15:03.924 } 00:15:03.924 ] 00:15:03.924 }' 00:15:03.924 16:03:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:03.924 16:03:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.184 [2024-11-20 16:03:02.223895] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:15:04.184 [2024-11-20 16:03:02.223952] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:04.184 [2024-11-20 16:03:02.223969] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:15:04.184 [2024-11-20 16:03:02.223980] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:04.184 [2024-11-20 16:03:02.224426] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:04.184 [2024-11-20 16:03:02.224444] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:15:04.184 [2024-11-20 16:03:02.224513] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:15:04.184 [2024-11-20 16:03:02.224537] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:04.184 [2024-11-20 16:03:02.224641] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:15:04.184 [2024-11-20 16:03:02.224652] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:15:04.184 [2024-11-20 16:03:02.224908] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:15:04.184 [2024-11-20 16:03:02.228386] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:15:04.184 [2024-11-20 16:03:02.228403] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:15:04.184 [2024-11-20 16:03:02.228645] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:04.184 pt3 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:04.184 "name": "raid_bdev1", 00:15:04.184 "uuid": "c8de4767-25b8-486a-8a7f-12aa6ee9cf59", 00:15:04.184 "strip_size_kb": 64, 00:15:04.184 "state": "online", 00:15:04.184 "raid_level": "raid5f", 00:15:04.184 "superblock": true, 00:15:04.184 "num_base_bdevs": 3, 00:15:04.184 "num_base_bdevs_discovered": 2, 00:15:04.184 "num_base_bdevs_operational": 2, 00:15:04.184 "base_bdevs_list": [ 00:15:04.184 { 00:15:04.184 "name": null, 00:15:04.184 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:04.184 "is_configured": false, 00:15:04.184 "data_offset": 2048, 00:15:04.184 "data_size": 63488 00:15:04.184 }, 00:15:04.184 { 00:15:04.184 "name": "pt2", 00:15:04.184 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:04.184 "is_configured": true, 00:15:04.184 "data_offset": 2048, 00:15:04.184 "data_size": 63488 00:15:04.184 }, 00:15:04.184 { 00:15:04.184 "name": "pt3", 00:15:04.184 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:04.184 "is_configured": true, 00:15:04.184 "data_offset": 2048, 00:15:04.184 "data_size": 63488 00:15:04.184 } 00:15:04.184 ] 00:15:04.184 }' 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:04.184 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.445 [2024-11-20 16:03:02.556700] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:04.445 [2024-11-20 16:03:02.556725] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:04.445 [2024-11-20 16:03:02.556788] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:04.445 [2024-11-20 16:03:02.556848] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:04.445 [2024-11-20 16:03:02.556857] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.445 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.445 [2024-11-20 16:03:02.608714] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:04.445 [2024-11-20 16:03:02.608761] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:04.445 [2024-11-20 16:03:02.608779] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:15:04.445 [2024-11-20 16:03:02.608787] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:04.445 [2024-11-20 16:03:02.611004] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:04.445 [2024-11-20 16:03:02.611148] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:04.445 [2024-11-20 16:03:02.611232] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:04.446 [2024-11-20 16:03:02.611276] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:04.446 [2024-11-20 16:03:02.611405] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:15:04.446 [2024-11-20 16:03:02.611415] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:04.446 [2024-11-20 16:03:02.611432] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:15:04.446 [2024-11-20 16:03:02.611477] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:04.446 pt1 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:04.446 "name": "raid_bdev1", 00:15:04.446 "uuid": "c8de4767-25b8-486a-8a7f-12aa6ee9cf59", 00:15:04.446 "strip_size_kb": 64, 00:15:04.446 "state": "configuring", 00:15:04.446 "raid_level": "raid5f", 00:15:04.446 "superblock": true, 00:15:04.446 "num_base_bdevs": 3, 00:15:04.446 "num_base_bdevs_discovered": 1, 00:15:04.446 "num_base_bdevs_operational": 2, 00:15:04.446 "base_bdevs_list": [ 00:15:04.446 { 00:15:04.446 "name": null, 00:15:04.446 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:04.446 "is_configured": false, 00:15:04.446 "data_offset": 2048, 00:15:04.446 "data_size": 63488 00:15:04.446 }, 00:15:04.446 { 00:15:04.446 "name": "pt2", 00:15:04.446 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:04.446 "is_configured": true, 00:15:04.446 "data_offset": 2048, 00:15:04.446 "data_size": 63488 00:15:04.446 }, 00:15:04.446 { 00:15:04.446 "name": null, 00:15:04.446 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:04.446 "is_configured": false, 00:15:04.446 "data_offset": 2048, 00:15:04.446 "data_size": 63488 00:15:04.446 } 00:15:04.446 ] 00:15:04.446 }' 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:04.446 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.705 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:04.705 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:15:04.705 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.705 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.705 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.966 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:15:04.966 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:15:04.966 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.966 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.966 [2024-11-20 16:03:02.964819] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:15:04.966 [2024-11-20 16:03:02.964876] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:04.966 [2024-11-20 16:03:02.964895] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:15:04.966 [2024-11-20 16:03:02.964904] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:04.966 [2024-11-20 16:03:02.965333] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:04.966 [2024-11-20 16:03:02.965349] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:15:04.966 [2024-11-20 16:03:02.965421] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:15:04.966 [2024-11-20 16:03:02.965441] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:04.966 [2024-11-20 16:03:02.965548] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:15:04.966 [2024-11-20 16:03:02.965557] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:15:04.966 [2024-11-20 16:03:02.965807] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:15:04.966 [2024-11-20 16:03:02.969489] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:15:04.966 [2024-11-20 16:03:02.969512] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:15:04.966 [2024-11-20 16:03:02.969754] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:04.966 pt3 00:15:04.966 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.966 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:04.966 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:04.966 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:04.966 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:04.966 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:04.966 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:04.966 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:04.967 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:04.967 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:04.967 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:04.967 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.967 16:03:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:04.967 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.967 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.967 16:03:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.967 16:03:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:04.967 "name": "raid_bdev1", 00:15:04.967 "uuid": "c8de4767-25b8-486a-8a7f-12aa6ee9cf59", 00:15:04.967 "strip_size_kb": 64, 00:15:04.967 "state": "online", 00:15:04.967 "raid_level": "raid5f", 00:15:04.967 "superblock": true, 00:15:04.967 "num_base_bdevs": 3, 00:15:04.967 "num_base_bdevs_discovered": 2, 00:15:04.967 "num_base_bdevs_operational": 2, 00:15:04.967 "base_bdevs_list": [ 00:15:04.967 { 00:15:04.967 "name": null, 00:15:04.967 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:04.967 "is_configured": false, 00:15:04.967 "data_offset": 2048, 00:15:04.967 "data_size": 63488 00:15:04.967 }, 00:15:04.967 { 00:15:04.967 "name": "pt2", 00:15:04.967 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:04.967 "is_configured": true, 00:15:04.967 "data_offset": 2048, 00:15:04.967 "data_size": 63488 00:15:04.967 }, 00:15:04.967 { 00:15:04.967 "name": "pt3", 00:15:04.967 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:04.967 "is_configured": true, 00:15:04.967 "data_offset": 2048, 00:15:04.967 "data_size": 63488 00:15:04.967 } 00:15:04.967 ] 00:15:04.967 }' 00:15:04.967 16:03:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:04.967 16:03:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.228 [2024-11-20 16:03:03.330072] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' c8de4767-25b8-486a-8a7f-12aa6ee9cf59 '!=' c8de4767-25b8-486a-8a7f-12aa6ee9cf59 ']' 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 78913 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 78913 ']' 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # kill -0 78913 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # uname 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 78913 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 78913' 00:15:05.228 killing process with pid 78913 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@973 -- # kill 78913 00:15:05.228 [2024-11-20 16:03:03.384560] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:05.228 16:03:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@978 -- # wait 78913 00:15:05.228 [2024-11-20 16:03:03.384685] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:05.228 [2024-11-20 16:03:03.384878] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:05.228 [2024-11-20 16:03:03.384921] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:15:05.486 [2024-11-20 16:03:03.571770] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:06.062 16:03:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:15:06.062 00:15:06.062 real 0m5.729s 00:15:06.062 user 0m8.951s 00:15:06.062 sys 0m0.893s 00:15:06.062 16:03:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:06.062 16:03:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.062 ************************************ 00:15:06.062 END TEST raid5f_superblock_test 00:15:06.062 ************************************ 00:15:06.324 16:03:04 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:15:06.324 16:03:04 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 3 false false true 00:15:06.324 16:03:04 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:15:06.324 16:03:04 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:06.324 16:03:04 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:06.324 ************************************ 00:15:06.324 START TEST raid5f_rebuild_test 00:15:06.324 ************************************ 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 3 false false true 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=79337 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 79337 00:15:06.324 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 79337 ']' 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.324 16:03:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:06.324 I/O size of 3145728 is greater than zero copy threshold (65536). 00:15:06.324 Zero copy mechanism will not be used. 00:15:06.324 [2024-11-20 16:03:04.424417] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:15:06.324 [2024-11-20 16:03:04.424541] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79337 ] 00:15:06.584 [2024-11-20 16:03:04.584036] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:06.584 [2024-11-20 16:03:04.687755] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:06.584 [2024-11-20 16:03:04.827973] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:06.584 [2024-11-20 16:03:04.828014] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.156 BaseBdev1_malloc 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.156 [2024-11-20 16:03:05.359572] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:07.156 [2024-11-20 16:03:05.359747] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:07.156 [2024-11-20 16:03:05.359775] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:07.156 [2024-11-20 16:03:05.359786] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:07.156 [2024-11-20 16:03:05.362012] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:07.156 [2024-11-20 16:03:05.362048] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:07.156 BaseBdev1 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.156 BaseBdev2_malloc 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.156 [2024-11-20 16:03:05.395824] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:07.156 [2024-11-20 16:03:05.395879] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:07.156 [2024-11-20 16:03:05.395900] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:07.156 [2024-11-20 16:03:05.395911] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:07.156 [2024-11-20 16:03:05.398032] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:07.156 [2024-11-20 16:03:05.398067] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:07.156 BaseBdev2 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.156 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.416 BaseBdev3_malloc 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.416 [2024-11-20 16:03:05.441217] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:15:07.416 [2024-11-20 16:03:05.441271] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:07.416 [2024-11-20 16:03:05.441292] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:15:07.416 [2024-11-20 16:03:05.441303] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:07.416 [2024-11-20 16:03:05.443466] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:07.416 [2024-11-20 16:03:05.443505] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:15:07.416 BaseBdev3 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.416 spare_malloc 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.416 spare_delay 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.416 [2024-11-20 16:03:05.485181] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:07.416 [2024-11-20 16:03:05.485230] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:07.416 [2024-11-20 16:03:05.485248] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:15:07.416 [2024-11-20 16:03:05.485259] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:07.416 [2024-11-20 16:03:05.487450] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:07.416 [2024-11-20 16:03:05.487492] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:07.416 spare 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.416 [2024-11-20 16:03:05.493245] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:07.416 [2024-11-20 16:03:05.495168] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:07.416 [2024-11-20 16:03:05.495233] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:07.416 [2024-11-20 16:03:05.495313] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:15:07.416 [2024-11-20 16:03:05.495324] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:15:07.416 [2024-11-20 16:03:05.495592] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:15:07.416 [2024-11-20 16:03:05.499408] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:15:07.416 [2024-11-20 16:03:05.499428] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:15:07.416 [2024-11-20 16:03:05.499599] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:07.416 "name": "raid_bdev1", 00:15:07.416 "uuid": "f313cae2-c293-42f8-8bf9-073095ddedc8", 00:15:07.416 "strip_size_kb": 64, 00:15:07.416 "state": "online", 00:15:07.416 "raid_level": "raid5f", 00:15:07.416 "superblock": false, 00:15:07.416 "num_base_bdevs": 3, 00:15:07.416 "num_base_bdevs_discovered": 3, 00:15:07.416 "num_base_bdevs_operational": 3, 00:15:07.416 "base_bdevs_list": [ 00:15:07.416 { 00:15:07.416 "name": "BaseBdev1", 00:15:07.416 "uuid": "bb786fdc-c2f9-513d-a47b-f877a5de1fc5", 00:15:07.416 "is_configured": true, 00:15:07.416 "data_offset": 0, 00:15:07.416 "data_size": 65536 00:15:07.416 }, 00:15:07.416 { 00:15:07.416 "name": "BaseBdev2", 00:15:07.416 "uuid": "1de585f0-42ce-5c73-92d9-c9c4e84aa986", 00:15:07.416 "is_configured": true, 00:15:07.416 "data_offset": 0, 00:15:07.416 "data_size": 65536 00:15:07.416 }, 00:15:07.416 { 00:15:07.416 "name": "BaseBdev3", 00:15:07.416 "uuid": "00c8a909-7aa1-5dfd-80ea-5cd66e8ef472", 00:15:07.416 "is_configured": true, 00:15:07.416 "data_offset": 0, 00:15:07.416 "data_size": 65536 00:15:07.416 } 00:15:07.416 ] 00:15:07.416 }' 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:07.416 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.678 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:07.678 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.678 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.678 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:07.678 [2024-11-20 16:03:05.864257] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:07.678 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.678 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=131072 00:15:07.678 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.678 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.678 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:07.678 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.678 16:03:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.939 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:15:07.939 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:07.939 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:07.939 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:07.939 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:07.939 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:07.939 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:07.939 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:07.939 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:07.939 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:07.939 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:15:07.939 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:07.939 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:07.939 16:03:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:07.939 [2024-11-20 16:03:06.116022] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:15:07.939 /dev/nbd0 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:07.939 1+0 records in 00:15:07.939 1+0 records out 00:15:07.939 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000684293 s, 6.0 MB/s 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 128 00:15:07.939 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=512 oflag=direct 00:15:08.509 512+0 records in 00:15:08.509 512+0 records out 00:15:08.509 67108864 bytes (67 MB, 64 MiB) copied, 0.514652 s, 130 MB/s 00:15:08.509 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:08.509 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:08.509 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:08.509 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:08.509 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:15:08.509 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:08.509 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:08.770 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:08.770 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:08.770 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:08.770 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:08.770 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:08.770 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:08.770 [2024-11-20 16:03:06.901327] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:08.770 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:15:08.770 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:15:08.770 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:08.770 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:08.770 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:08.770 [2024-11-20 16:03:06.909421] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:08.770 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:08.770 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:08.770 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:08.771 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:08.771 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:08.771 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:08.771 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:08.771 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:08.771 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:08.771 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:08.771 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:08.771 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:08.771 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:08.771 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:08.771 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:08.771 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:08.771 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:08.771 "name": "raid_bdev1", 00:15:08.771 "uuid": "f313cae2-c293-42f8-8bf9-073095ddedc8", 00:15:08.771 "strip_size_kb": 64, 00:15:08.771 "state": "online", 00:15:08.771 "raid_level": "raid5f", 00:15:08.771 "superblock": false, 00:15:08.771 "num_base_bdevs": 3, 00:15:08.771 "num_base_bdevs_discovered": 2, 00:15:08.771 "num_base_bdevs_operational": 2, 00:15:08.771 "base_bdevs_list": [ 00:15:08.771 { 00:15:08.771 "name": null, 00:15:08.771 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:08.771 "is_configured": false, 00:15:08.771 "data_offset": 0, 00:15:08.771 "data_size": 65536 00:15:08.771 }, 00:15:08.771 { 00:15:08.771 "name": "BaseBdev2", 00:15:08.771 "uuid": "1de585f0-42ce-5c73-92d9-c9c4e84aa986", 00:15:08.771 "is_configured": true, 00:15:08.771 "data_offset": 0, 00:15:08.771 "data_size": 65536 00:15:08.771 }, 00:15:08.771 { 00:15:08.771 "name": "BaseBdev3", 00:15:08.771 "uuid": "00c8a909-7aa1-5dfd-80ea-5cd66e8ef472", 00:15:08.771 "is_configured": true, 00:15:08.771 "data_offset": 0, 00:15:08.771 "data_size": 65536 00:15:08.771 } 00:15:08.771 ] 00:15:08.771 }' 00:15:08.771 16:03:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:08.771 16:03:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:09.096 16:03:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:09.096 16:03:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.096 16:03:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:09.096 [2024-11-20 16:03:07.225503] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:09.096 [2024-11-20 16:03:07.236351] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b680 00:15:09.096 16:03:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.096 16:03:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:09.096 [2024-11-20 16:03:07.241924] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:10.038 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:10.038 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:10.038 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:10.038 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:10.038 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:10.038 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:10.038 16:03:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.038 16:03:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.038 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:10.038 16:03:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.038 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:10.038 "name": "raid_bdev1", 00:15:10.038 "uuid": "f313cae2-c293-42f8-8bf9-073095ddedc8", 00:15:10.038 "strip_size_kb": 64, 00:15:10.038 "state": "online", 00:15:10.038 "raid_level": "raid5f", 00:15:10.038 "superblock": false, 00:15:10.038 "num_base_bdevs": 3, 00:15:10.038 "num_base_bdevs_discovered": 3, 00:15:10.038 "num_base_bdevs_operational": 3, 00:15:10.038 "process": { 00:15:10.038 "type": "rebuild", 00:15:10.038 "target": "spare", 00:15:10.038 "progress": { 00:15:10.038 "blocks": 18432, 00:15:10.038 "percent": 14 00:15:10.038 } 00:15:10.038 }, 00:15:10.038 "base_bdevs_list": [ 00:15:10.038 { 00:15:10.038 "name": "spare", 00:15:10.038 "uuid": "20f3475b-6dee-5258-8ea3-5c13305b6838", 00:15:10.038 "is_configured": true, 00:15:10.038 "data_offset": 0, 00:15:10.038 "data_size": 65536 00:15:10.039 }, 00:15:10.039 { 00:15:10.039 "name": "BaseBdev2", 00:15:10.039 "uuid": "1de585f0-42ce-5c73-92d9-c9c4e84aa986", 00:15:10.039 "is_configured": true, 00:15:10.039 "data_offset": 0, 00:15:10.039 "data_size": 65536 00:15:10.039 }, 00:15:10.039 { 00:15:10.039 "name": "BaseBdev3", 00:15:10.039 "uuid": "00c8a909-7aa1-5dfd-80ea-5cd66e8ef472", 00:15:10.039 "is_configured": true, 00:15:10.039 "data_offset": 0, 00:15:10.039 "data_size": 65536 00:15:10.039 } 00:15:10.039 ] 00:15:10.039 }' 00:15:10.039 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.298 [2024-11-20 16:03:08.351138] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:10.298 [2024-11-20 16:03:08.352316] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:10.298 [2024-11-20 16:03:08.352369] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:10.298 [2024-11-20 16:03:08.352387] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:10.298 [2024-11-20 16:03:08.352394] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:10.298 "name": "raid_bdev1", 00:15:10.298 "uuid": "f313cae2-c293-42f8-8bf9-073095ddedc8", 00:15:10.298 "strip_size_kb": 64, 00:15:10.298 "state": "online", 00:15:10.298 "raid_level": "raid5f", 00:15:10.298 "superblock": false, 00:15:10.298 "num_base_bdevs": 3, 00:15:10.298 "num_base_bdevs_discovered": 2, 00:15:10.298 "num_base_bdevs_operational": 2, 00:15:10.298 "base_bdevs_list": [ 00:15:10.298 { 00:15:10.298 "name": null, 00:15:10.298 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:10.298 "is_configured": false, 00:15:10.298 "data_offset": 0, 00:15:10.298 "data_size": 65536 00:15:10.298 }, 00:15:10.298 { 00:15:10.298 "name": "BaseBdev2", 00:15:10.298 "uuid": "1de585f0-42ce-5c73-92d9-c9c4e84aa986", 00:15:10.298 "is_configured": true, 00:15:10.298 "data_offset": 0, 00:15:10.298 "data_size": 65536 00:15:10.298 }, 00:15:10.298 { 00:15:10.298 "name": "BaseBdev3", 00:15:10.298 "uuid": "00c8a909-7aa1-5dfd-80ea-5cd66e8ef472", 00:15:10.298 "is_configured": true, 00:15:10.298 "data_offset": 0, 00:15:10.298 "data_size": 65536 00:15:10.298 } 00:15:10.298 ] 00:15:10.298 }' 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:10.298 16:03:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:10.560 "name": "raid_bdev1", 00:15:10.560 "uuid": "f313cae2-c293-42f8-8bf9-073095ddedc8", 00:15:10.560 "strip_size_kb": 64, 00:15:10.560 "state": "online", 00:15:10.560 "raid_level": "raid5f", 00:15:10.560 "superblock": false, 00:15:10.560 "num_base_bdevs": 3, 00:15:10.560 "num_base_bdevs_discovered": 2, 00:15:10.560 "num_base_bdevs_operational": 2, 00:15:10.560 "base_bdevs_list": [ 00:15:10.560 { 00:15:10.560 "name": null, 00:15:10.560 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:10.560 "is_configured": false, 00:15:10.560 "data_offset": 0, 00:15:10.560 "data_size": 65536 00:15:10.560 }, 00:15:10.560 { 00:15:10.560 "name": "BaseBdev2", 00:15:10.560 "uuid": "1de585f0-42ce-5c73-92d9-c9c4e84aa986", 00:15:10.560 "is_configured": true, 00:15:10.560 "data_offset": 0, 00:15:10.560 "data_size": 65536 00:15:10.560 }, 00:15:10.560 { 00:15:10.560 "name": "BaseBdev3", 00:15:10.560 "uuid": "00c8a909-7aa1-5dfd-80ea-5cd66e8ef472", 00:15:10.560 "is_configured": true, 00:15:10.560 "data_offset": 0, 00:15:10.560 "data_size": 65536 00:15:10.560 } 00:15:10.560 ] 00:15:10.560 }' 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.560 16:03:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.560 [2024-11-20 16:03:08.802816] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:10.822 [2024-11-20 16:03:08.812761] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b750 00:15:10.822 16:03:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.822 16:03:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:10.822 [2024-11-20 16:03:08.818068] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:11.763 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:11.763 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:11.763 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:11.763 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:11.763 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:11.763 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.763 16:03:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.763 16:03:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.763 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:11.763 16:03:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.763 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:11.763 "name": "raid_bdev1", 00:15:11.763 "uuid": "f313cae2-c293-42f8-8bf9-073095ddedc8", 00:15:11.763 "strip_size_kb": 64, 00:15:11.763 "state": "online", 00:15:11.763 "raid_level": "raid5f", 00:15:11.763 "superblock": false, 00:15:11.763 "num_base_bdevs": 3, 00:15:11.763 "num_base_bdevs_discovered": 3, 00:15:11.763 "num_base_bdevs_operational": 3, 00:15:11.763 "process": { 00:15:11.763 "type": "rebuild", 00:15:11.763 "target": "spare", 00:15:11.763 "progress": { 00:15:11.763 "blocks": 18432, 00:15:11.763 "percent": 14 00:15:11.763 } 00:15:11.763 }, 00:15:11.763 "base_bdevs_list": [ 00:15:11.763 { 00:15:11.763 "name": "spare", 00:15:11.763 "uuid": "20f3475b-6dee-5258-8ea3-5c13305b6838", 00:15:11.763 "is_configured": true, 00:15:11.763 "data_offset": 0, 00:15:11.763 "data_size": 65536 00:15:11.763 }, 00:15:11.763 { 00:15:11.763 "name": "BaseBdev2", 00:15:11.763 "uuid": "1de585f0-42ce-5c73-92d9-c9c4e84aa986", 00:15:11.763 "is_configured": true, 00:15:11.763 "data_offset": 0, 00:15:11.763 "data_size": 65536 00:15:11.763 }, 00:15:11.763 { 00:15:11.763 "name": "BaseBdev3", 00:15:11.763 "uuid": "00c8a909-7aa1-5dfd-80ea-5cd66e8ef472", 00:15:11.763 "is_configured": true, 00:15:11.763 "data_offset": 0, 00:15:11.763 "data_size": 65536 00:15:11.763 } 00:15:11.763 ] 00:15:11.763 }' 00:15:11.763 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=443 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:11.764 "name": "raid_bdev1", 00:15:11.764 "uuid": "f313cae2-c293-42f8-8bf9-073095ddedc8", 00:15:11.764 "strip_size_kb": 64, 00:15:11.764 "state": "online", 00:15:11.764 "raid_level": "raid5f", 00:15:11.764 "superblock": false, 00:15:11.764 "num_base_bdevs": 3, 00:15:11.764 "num_base_bdevs_discovered": 3, 00:15:11.764 "num_base_bdevs_operational": 3, 00:15:11.764 "process": { 00:15:11.764 "type": "rebuild", 00:15:11.764 "target": "spare", 00:15:11.764 "progress": { 00:15:11.764 "blocks": 20480, 00:15:11.764 "percent": 15 00:15:11.764 } 00:15:11.764 }, 00:15:11.764 "base_bdevs_list": [ 00:15:11.764 { 00:15:11.764 "name": "spare", 00:15:11.764 "uuid": "20f3475b-6dee-5258-8ea3-5c13305b6838", 00:15:11.764 "is_configured": true, 00:15:11.764 "data_offset": 0, 00:15:11.764 "data_size": 65536 00:15:11.764 }, 00:15:11.764 { 00:15:11.764 "name": "BaseBdev2", 00:15:11.764 "uuid": "1de585f0-42ce-5c73-92d9-c9c4e84aa986", 00:15:11.764 "is_configured": true, 00:15:11.764 "data_offset": 0, 00:15:11.764 "data_size": 65536 00:15:11.764 }, 00:15:11.764 { 00:15:11.764 "name": "BaseBdev3", 00:15:11.764 "uuid": "00c8a909-7aa1-5dfd-80ea-5cd66e8ef472", 00:15:11.764 "is_configured": true, 00:15:11.764 "data_offset": 0, 00:15:11.764 "data_size": 65536 00:15:11.764 } 00:15:11.764 ] 00:15:11.764 }' 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:11.764 16:03:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:12.024 16:03:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:12.024 16:03:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:12.963 16:03:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:12.964 16:03:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:12.964 16:03:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:12.964 16:03:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:12.964 16:03:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:12.964 16:03:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:12.964 16:03:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:12.964 16:03:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:12.964 16:03:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:12.964 16:03:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:12.964 16:03:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:12.964 16:03:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:12.964 "name": "raid_bdev1", 00:15:12.964 "uuid": "f313cae2-c293-42f8-8bf9-073095ddedc8", 00:15:12.964 "strip_size_kb": 64, 00:15:12.964 "state": "online", 00:15:12.964 "raid_level": "raid5f", 00:15:12.964 "superblock": false, 00:15:12.964 "num_base_bdevs": 3, 00:15:12.964 "num_base_bdevs_discovered": 3, 00:15:12.964 "num_base_bdevs_operational": 3, 00:15:12.964 "process": { 00:15:12.964 "type": "rebuild", 00:15:12.964 "target": "spare", 00:15:12.964 "progress": { 00:15:12.964 "blocks": 45056, 00:15:12.964 "percent": 34 00:15:12.964 } 00:15:12.964 }, 00:15:12.964 "base_bdevs_list": [ 00:15:12.964 { 00:15:12.964 "name": "spare", 00:15:12.964 "uuid": "20f3475b-6dee-5258-8ea3-5c13305b6838", 00:15:12.964 "is_configured": true, 00:15:12.964 "data_offset": 0, 00:15:12.964 "data_size": 65536 00:15:12.964 }, 00:15:12.964 { 00:15:12.964 "name": "BaseBdev2", 00:15:12.964 "uuid": "1de585f0-42ce-5c73-92d9-c9c4e84aa986", 00:15:12.964 "is_configured": true, 00:15:12.964 "data_offset": 0, 00:15:12.964 "data_size": 65536 00:15:12.964 }, 00:15:12.964 { 00:15:12.964 "name": "BaseBdev3", 00:15:12.964 "uuid": "00c8a909-7aa1-5dfd-80ea-5cd66e8ef472", 00:15:12.964 "is_configured": true, 00:15:12.964 "data_offset": 0, 00:15:12.964 "data_size": 65536 00:15:12.964 } 00:15:12.964 ] 00:15:12.964 }' 00:15:12.964 16:03:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:12.964 16:03:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:12.964 16:03:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:12.964 16:03:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:12.964 16:03:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:13.905 16:03:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:13.905 16:03:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:13.905 16:03:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:13.905 16:03:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:13.905 16:03:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:13.905 16:03:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:13.905 16:03:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.905 16:03:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:13.905 16:03:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.905 16:03:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:14.165 16:03:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.165 16:03:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:14.165 "name": "raid_bdev1", 00:15:14.165 "uuid": "f313cae2-c293-42f8-8bf9-073095ddedc8", 00:15:14.165 "strip_size_kb": 64, 00:15:14.165 "state": "online", 00:15:14.165 "raid_level": "raid5f", 00:15:14.165 "superblock": false, 00:15:14.165 "num_base_bdevs": 3, 00:15:14.165 "num_base_bdevs_discovered": 3, 00:15:14.165 "num_base_bdevs_operational": 3, 00:15:14.165 "process": { 00:15:14.165 "type": "rebuild", 00:15:14.165 "target": "spare", 00:15:14.165 "progress": { 00:15:14.165 "blocks": 67584, 00:15:14.165 "percent": 51 00:15:14.165 } 00:15:14.165 }, 00:15:14.165 "base_bdevs_list": [ 00:15:14.165 { 00:15:14.165 "name": "spare", 00:15:14.165 "uuid": "20f3475b-6dee-5258-8ea3-5c13305b6838", 00:15:14.165 "is_configured": true, 00:15:14.165 "data_offset": 0, 00:15:14.165 "data_size": 65536 00:15:14.165 }, 00:15:14.165 { 00:15:14.165 "name": "BaseBdev2", 00:15:14.165 "uuid": "1de585f0-42ce-5c73-92d9-c9c4e84aa986", 00:15:14.165 "is_configured": true, 00:15:14.165 "data_offset": 0, 00:15:14.165 "data_size": 65536 00:15:14.165 }, 00:15:14.165 { 00:15:14.165 "name": "BaseBdev3", 00:15:14.165 "uuid": "00c8a909-7aa1-5dfd-80ea-5cd66e8ef472", 00:15:14.165 "is_configured": true, 00:15:14.165 "data_offset": 0, 00:15:14.165 "data_size": 65536 00:15:14.165 } 00:15:14.165 ] 00:15:14.165 }' 00:15:14.165 16:03:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:14.165 16:03:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:14.165 16:03:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:14.165 16:03:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:14.165 16:03:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:15.104 16:03:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:15.105 16:03:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:15.105 16:03:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:15.105 16:03:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:15.105 16:03:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:15.105 16:03:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:15.105 16:03:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.105 16:03:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.105 16:03:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.105 16:03:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:15.105 16:03:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.105 16:03:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:15.105 "name": "raid_bdev1", 00:15:15.105 "uuid": "f313cae2-c293-42f8-8bf9-073095ddedc8", 00:15:15.105 "strip_size_kb": 64, 00:15:15.105 "state": "online", 00:15:15.105 "raid_level": "raid5f", 00:15:15.105 "superblock": false, 00:15:15.105 "num_base_bdevs": 3, 00:15:15.105 "num_base_bdevs_discovered": 3, 00:15:15.105 "num_base_bdevs_operational": 3, 00:15:15.105 "process": { 00:15:15.105 "type": "rebuild", 00:15:15.105 "target": "spare", 00:15:15.105 "progress": { 00:15:15.105 "blocks": 88064, 00:15:15.105 "percent": 67 00:15:15.105 } 00:15:15.105 }, 00:15:15.105 "base_bdevs_list": [ 00:15:15.105 { 00:15:15.105 "name": "spare", 00:15:15.105 "uuid": "20f3475b-6dee-5258-8ea3-5c13305b6838", 00:15:15.105 "is_configured": true, 00:15:15.105 "data_offset": 0, 00:15:15.105 "data_size": 65536 00:15:15.105 }, 00:15:15.105 { 00:15:15.105 "name": "BaseBdev2", 00:15:15.105 "uuid": "1de585f0-42ce-5c73-92d9-c9c4e84aa986", 00:15:15.105 "is_configured": true, 00:15:15.105 "data_offset": 0, 00:15:15.105 "data_size": 65536 00:15:15.105 }, 00:15:15.105 { 00:15:15.105 "name": "BaseBdev3", 00:15:15.105 "uuid": "00c8a909-7aa1-5dfd-80ea-5cd66e8ef472", 00:15:15.105 "is_configured": true, 00:15:15.105 "data_offset": 0, 00:15:15.105 "data_size": 65536 00:15:15.105 } 00:15:15.105 ] 00:15:15.105 }' 00:15:15.105 16:03:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:15.105 16:03:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:15.105 16:03:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:15.105 16:03:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:15.105 16:03:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:16.558 16:03:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:16.558 16:03:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:16.558 16:03:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:16.558 16:03:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:16.558 16:03:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:16.558 16:03:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:16.558 16:03:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:16.558 16:03:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:16.558 16:03:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.558 16:03:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:16.558 16:03:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.558 16:03:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:16.558 "name": "raid_bdev1", 00:15:16.558 "uuid": "f313cae2-c293-42f8-8bf9-073095ddedc8", 00:15:16.558 "strip_size_kb": 64, 00:15:16.558 "state": "online", 00:15:16.558 "raid_level": "raid5f", 00:15:16.558 "superblock": false, 00:15:16.558 "num_base_bdevs": 3, 00:15:16.558 "num_base_bdevs_discovered": 3, 00:15:16.558 "num_base_bdevs_operational": 3, 00:15:16.558 "process": { 00:15:16.558 "type": "rebuild", 00:15:16.558 "target": "spare", 00:15:16.558 "progress": { 00:15:16.558 "blocks": 110592, 00:15:16.558 "percent": 84 00:15:16.558 } 00:15:16.558 }, 00:15:16.558 "base_bdevs_list": [ 00:15:16.558 { 00:15:16.558 "name": "spare", 00:15:16.558 "uuid": "20f3475b-6dee-5258-8ea3-5c13305b6838", 00:15:16.558 "is_configured": true, 00:15:16.558 "data_offset": 0, 00:15:16.558 "data_size": 65536 00:15:16.558 }, 00:15:16.558 { 00:15:16.558 "name": "BaseBdev2", 00:15:16.558 "uuid": "1de585f0-42ce-5c73-92d9-c9c4e84aa986", 00:15:16.558 "is_configured": true, 00:15:16.558 "data_offset": 0, 00:15:16.558 "data_size": 65536 00:15:16.558 }, 00:15:16.558 { 00:15:16.558 "name": "BaseBdev3", 00:15:16.558 "uuid": "00c8a909-7aa1-5dfd-80ea-5cd66e8ef472", 00:15:16.558 "is_configured": true, 00:15:16.558 "data_offset": 0, 00:15:16.558 "data_size": 65536 00:15:16.558 } 00:15:16.558 ] 00:15:16.558 }' 00:15:16.558 16:03:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:16.558 16:03:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:16.558 16:03:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:16.558 16:03:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:16.558 16:03:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:17.128 [2024-11-20 16:03:15.277119] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:17.128 [2024-11-20 16:03:15.277362] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:17.128 [2024-11-20 16:03:15.277413] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:17.387 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:17.387 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:17.387 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:17.387 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:17.387 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:17.387 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:17.387 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:17.387 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:17.388 "name": "raid_bdev1", 00:15:17.388 "uuid": "f313cae2-c293-42f8-8bf9-073095ddedc8", 00:15:17.388 "strip_size_kb": 64, 00:15:17.388 "state": "online", 00:15:17.388 "raid_level": "raid5f", 00:15:17.388 "superblock": false, 00:15:17.388 "num_base_bdevs": 3, 00:15:17.388 "num_base_bdevs_discovered": 3, 00:15:17.388 "num_base_bdevs_operational": 3, 00:15:17.388 "base_bdevs_list": [ 00:15:17.388 { 00:15:17.388 "name": "spare", 00:15:17.388 "uuid": "20f3475b-6dee-5258-8ea3-5c13305b6838", 00:15:17.388 "is_configured": true, 00:15:17.388 "data_offset": 0, 00:15:17.388 "data_size": 65536 00:15:17.388 }, 00:15:17.388 { 00:15:17.388 "name": "BaseBdev2", 00:15:17.388 "uuid": "1de585f0-42ce-5c73-92d9-c9c4e84aa986", 00:15:17.388 "is_configured": true, 00:15:17.388 "data_offset": 0, 00:15:17.388 "data_size": 65536 00:15:17.388 }, 00:15:17.388 { 00:15:17.388 "name": "BaseBdev3", 00:15:17.388 "uuid": "00c8a909-7aa1-5dfd-80ea-5cd66e8ef472", 00:15:17.388 "is_configured": true, 00:15:17.388 "data_offset": 0, 00:15:17.388 "data_size": 65536 00:15:17.388 } 00:15:17.388 ] 00:15:17.388 }' 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:17.388 "name": "raid_bdev1", 00:15:17.388 "uuid": "f313cae2-c293-42f8-8bf9-073095ddedc8", 00:15:17.388 "strip_size_kb": 64, 00:15:17.388 "state": "online", 00:15:17.388 "raid_level": "raid5f", 00:15:17.388 "superblock": false, 00:15:17.388 "num_base_bdevs": 3, 00:15:17.388 "num_base_bdevs_discovered": 3, 00:15:17.388 "num_base_bdevs_operational": 3, 00:15:17.388 "base_bdevs_list": [ 00:15:17.388 { 00:15:17.388 "name": "spare", 00:15:17.388 "uuid": "20f3475b-6dee-5258-8ea3-5c13305b6838", 00:15:17.388 "is_configured": true, 00:15:17.388 "data_offset": 0, 00:15:17.388 "data_size": 65536 00:15:17.388 }, 00:15:17.388 { 00:15:17.388 "name": "BaseBdev2", 00:15:17.388 "uuid": "1de585f0-42ce-5c73-92d9-c9c4e84aa986", 00:15:17.388 "is_configured": true, 00:15:17.388 "data_offset": 0, 00:15:17.388 "data_size": 65536 00:15:17.388 }, 00:15:17.388 { 00:15:17.388 "name": "BaseBdev3", 00:15:17.388 "uuid": "00c8a909-7aa1-5dfd-80ea-5cd66e8ef472", 00:15:17.388 "is_configured": true, 00:15:17.388 "data_offset": 0, 00:15:17.388 "data_size": 65536 00:15:17.388 } 00:15:17.388 ] 00:15:17.388 }' 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:17.388 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:17.647 "name": "raid_bdev1", 00:15:17.647 "uuid": "f313cae2-c293-42f8-8bf9-073095ddedc8", 00:15:17.647 "strip_size_kb": 64, 00:15:17.647 "state": "online", 00:15:17.647 "raid_level": "raid5f", 00:15:17.647 "superblock": false, 00:15:17.647 "num_base_bdevs": 3, 00:15:17.647 "num_base_bdevs_discovered": 3, 00:15:17.647 "num_base_bdevs_operational": 3, 00:15:17.647 "base_bdevs_list": [ 00:15:17.647 { 00:15:17.647 "name": "spare", 00:15:17.647 "uuid": "20f3475b-6dee-5258-8ea3-5c13305b6838", 00:15:17.647 "is_configured": true, 00:15:17.647 "data_offset": 0, 00:15:17.647 "data_size": 65536 00:15:17.647 }, 00:15:17.647 { 00:15:17.647 "name": "BaseBdev2", 00:15:17.647 "uuid": "1de585f0-42ce-5c73-92d9-c9c4e84aa986", 00:15:17.647 "is_configured": true, 00:15:17.647 "data_offset": 0, 00:15:17.647 "data_size": 65536 00:15:17.647 }, 00:15:17.647 { 00:15:17.647 "name": "BaseBdev3", 00:15:17.647 "uuid": "00c8a909-7aa1-5dfd-80ea-5cd66e8ef472", 00:15:17.647 "is_configured": true, 00:15:17.647 "data_offset": 0, 00:15:17.647 "data_size": 65536 00:15:17.647 } 00:15:17.647 ] 00:15:17.647 }' 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:17.647 16:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.907 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:17.907 16:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:17.907 16:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.907 [2024-11-20 16:03:15.979416] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:17.907 [2024-11-20 16:03:15.979538] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:17.907 [2024-11-20 16:03:15.979623] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:17.907 [2024-11-20 16:03:15.979726] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:17.907 [2024-11-20 16:03:15.979743] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:15:17.907 16:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:17.907 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:15:17.907 16:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:17.907 16:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:17.907 16:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.907 16:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:17.907 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:17.907 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:17.907 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:17.907 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:17.907 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:17.907 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:17.907 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:17.907 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:17.907 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:17.907 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:15:17.907 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:17.907 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:17.907 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:18.167 /dev/nbd0 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:18.167 1+0 records in 00:15:18.167 1+0 records out 00:15:18.167 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000211415 s, 19.4 MB/s 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:18.167 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:18.429 /dev/nbd1 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:18.430 1+0 records in 00:15:18.430 1+0 records out 00:15:18.430 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000511786 s, 8.0 MB/s 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:18.430 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:18.691 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:18.691 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:18.691 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:18.691 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:18.691 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:18.691 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:18.691 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:15:18.691 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:15:18.691 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:18.691 16:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 79337 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 79337 ']' 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 79337 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79337 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:18.953 killing process with pid 79337 00:15:18.953 Received shutdown signal, test time was about 60.000000 seconds 00:15:18.953 00:15:18.953 Latency(us) 00:15:18.953 [2024-11-20T16:03:17.204Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:18.953 [2024-11-20T16:03:17.204Z] =================================================================================================================== 00:15:18.953 [2024-11-20T16:03:17.204Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79337' 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@973 -- # kill 79337 00:15:18.953 [2024-11-20 16:03:17.129695] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:18.953 16:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@978 -- # wait 79337 00:15:19.215 [2024-11-20 16:03:17.373177] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:15:20.155 00:15:20.155 real 0m13.812s 00:15:20.155 user 0m16.704s 00:15:20.155 sys 0m1.605s 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:20.155 ************************************ 00:15:20.155 END TEST raid5f_rebuild_test 00:15:20.155 ************************************ 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:20.155 16:03:18 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 3 true false true 00:15:20.155 16:03:18 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:15:20.155 16:03:18 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:20.155 16:03:18 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:20.155 ************************************ 00:15:20.155 START TEST raid5f_rebuild_test_sb 00:15:20.155 ************************************ 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 3 true false true 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=79759 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 79759 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 79759 ']' 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:20.155 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:20.155 16:03:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:20.155 I/O size of 3145728 is greater than zero copy threshold (65536). 00:15:20.155 Zero copy mechanism will not be used. 00:15:20.155 [2024-11-20 16:03:18.321057] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:15:20.155 [2024-11-20 16:03:18.321223] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79759 ] 00:15:20.415 [2024-11-20 16:03:18.488556] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:20.415 [2024-11-20 16:03:18.590191] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:20.675 [2024-11-20 16:03:18.726772] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:20.675 [2024-11-20 16:03:18.726824] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:20.934 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:20.934 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:15:20.934 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:20.934 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:15:20.934 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:20.934 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:20.934 BaseBdev1_malloc 00:15:20.934 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:20.934 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:20.934 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:20.934 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:20.934 [2024-11-20 16:03:19.181515] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:20.934 [2024-11-20 16:03:19.181574] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:20.934 [2024-11-20 16:03:19.181595] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:20.934 [2024-11-20 16:03:19.181606] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:21.193 [2024-11-20 16:03:19.183778] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:21.193 [2024-11-20 16:03:19.183812] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:21.193 BaseBdev1 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.193 BaseBdev2_malloc 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.193 [2024-11-20 16:03:19.217721] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:21.193 [2024-11-20 16:03:19.217772] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:21.193 [2024-11-20 16:03:19.217794] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:21.193 [2024-11-20 16:03:19.217806] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:21.193 [2024-11-20 16:03:19.219909] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:21.193 [2024-11-20 16:03:19.219941] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:21.193 BaseBdev2 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.193 BaseBdev3_malloc 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.193 [2024-11-20 16:03:19.264813] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:15:21.193 [2024-11-20 16:03:19.264871] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:21.193 [2024-11-20 16:03:19.264894] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:15:21.193 [2024-11-20 16:03:19.264905] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:21.193 [2024-11-20 16:03:19.267049] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:21.193 [2024-11-20 16:03:19.267199] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:15:21.193 BaseBdev3 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.193 spare_malloc 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.193 spare_delay 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.193 [2024-11-20 16:03:19.308849] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:21.193 [2024-11-20 16:03:19.308895] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:21.193 [2024-11-20 16:03:19.308912] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:15:21.193 [2024-11-20 16:03:19.308924] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:21.193 [2024-11-20 16:03:19.311034] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:21.193 [2024-11-20 16:03:19.311070] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:21.193 spare 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.193 [2024-11-20 16:03:19.316915] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:21.193 [2024-11-20 16:03:19.318749] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:21.193 [2024-11-20 16:03:19.318813] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:21.193 [2024-11-20 16:03:19.318980] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:15:21.193 [2024-11-20 16:03:19.318991] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:15:21.193 [2024-11-20 16:03:19.319240] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:15:21.193 [2024-11-20 16:03:19.323002] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:15:21.193 [2024-11-20 16:03:19.323023] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:15:21.193 [2024-11-20 16:03:19.323192] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:21.193 "name": "raid_bdev1", 00:15:21.193 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:21.193 "strip_size_kb": 64, 00:15:21.193 "state": "online", 00:15:21.193 "raid_level": "raid5f", 00:15:21.193 "superblock": true, 00:15:21.193 "num_base_bdevs": 3, 00:15:21.193 "num_base_bdevs_discovered": 3, 00:15:21.193 "num_base_bdevs_operational": 3, 00:15:21.193 "base_bdevs_list": [ 00:15:21.193 { 00:15:21.193 "name": "BaseBdev1", 00:15:21.193 "uuid": "781c9f5d-49ec-5b0c-89b2-470959e9ee69", 00:15:21.193 "is_configured": true, 00:15:21.193 "data_offset": 2048, 00:15:21.193 "data_size": 63488 00:15:21.193 }, 00:15:21.193 { 00:15:21.193 "name": "BaseBdev2", 00:15:21.193 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:21.193 "is_configured": true, 00:15:21.193 "data_offset": 2048, 00:15:21.193 "data_size": 63488 00:15:21.193 }, 00:15:21.193 { 00:15:21.193 "name": "BaseBdev3", 00:15:21.193 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:21.193 "is_configured": true, 00:15:21.193 "data_offset": 2048, 00:15:21.193 "data_size": 63488 00:15:21.193 } 00:15:21.193 ] 00:15:21.193 }' 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:21.193 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.454 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:21.454 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:21.454 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.454 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.454 [2024-11-20 16:03:19.675499] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:21.454 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.454 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=126976 00:15:21.454 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:21.454 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:21.454 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.454 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:21.714 [2024-11-20 16:03:19.919396] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:15:21.714 /dev/nbd0 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:21.714 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:21.715 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:21.975 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:15:21.975 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:21.975 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:21.975 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:21.975 1+0 records in 00:15:21.975 1+0 records out 00:15:21.975 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00050181 s, 8.2 MB/s 00:15:21.975 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:21.975 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:15:21.975 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:21.975 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:21.975 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:15:21.975 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:21.975 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:21.975 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:15:21.975 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:15:21.975 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 128 00:15:21.975 16:03:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=496 oflag=direct 00:15:22.545 496+0 records in 00:15:22.545 496+0 records out 00:15:22.545 65011712 bytes (65 MB, 62 MiB) copied, 0.512855 s, 127 MB/s 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:22.545 [2024-11-20 16:03:20.752150] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:22.545 [2024-11-20 16:03:20.764227] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:22.545 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:22.804 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:22.804 "name": "raid_bdev1", 00:15:22.804 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:22.804 "strip_size_kb": 64, 00:15:22.804 "state": "online", 00:15:22.804 "raid_level": "raid5f", 00:15:22.804 "superblock": true, 00:15:22.804 "num_base_bdevs": 3, 00:15:22.804 "num_base_bdevs_discovered": 2, 00:15:22.804 "num_base_bdevs_operational": 2, 00:15:22.804 "base_bdevs_list": [ 00:15:22.804 { 00:15:22.804 "name": null, 00:15:22.804 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:22.804 "is_configured": false, 00:15:22.804 "data_offset": 0, 00:15:22.804 "data_size": 63488 00:15:22.804 }, 00:15:22.804 { 00:15:22.804 "name": "BaseBdev2", 00:15:22.804 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:22.804 "is_configured": true, 00:15:22.804 "data_offset": 2048, 00:15:22.804 "data_size": 63488 00:15:22.804 }, 00:15:22.804 { 00:15:22.804 "name": "BaseBdev3", 00:15:22.804 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:22.804 "is_configured": true, 00:15:22.804 "data_offset": 2048, 00:15:22.804 "data_size": 63488 00:15:22.804 } 00:15:22.804 ] 00:15:22.804 }' 00:15:22.804 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:22.804 16:03:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:23.064 16:03:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:23.064 16:03:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:23.064 16:03:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:23.064 [2024-11-20 16:03:21.104313] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:23.064 [2024-11-20 16:03:21.115121] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000028f80 00:15:23.064 16:03:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:23.064 16:03:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:23.064 [2024-11-20 16:03:21.120586] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:24.004 "name": "raid_bdev1", 00:15:24.004 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:24.004 "strip_size_kb": 64, 00:15:24.004 "state": "online", 00:15:24.004 "raid_level": "raid5f", 00:15:24.004 "superblock": true, 00:15:24.004 "num_base_bdevs": 3, 00:15:24.004 "num_base_bdevs_discovered": 3, 00:15:24.004 "num_base_bdevs_operational": 3, 00:15:24.004 "process": { 00:15:24.004 "type": "rebuild", 00:15:24.004 "target": "spare", 00:15:24.004 "progress": { 00:15:24.004 "blocks": 18432, 00:15:24.004 "percent": 14 00:15:24.004 } 00:15:24.004 }, 00:15:24.004 "base_bdevs_list": [ 00:15:24.004 { 00:15:24.004 "name": "spare", 00:15:24.004 "uuid": "c3caf461-29a0-503c-b7d4-35b7fd697eec", 00:15:24.004 "is_configured": true, 00:15:24.004 "data_offset": 2048, 00:15:24.004 "data_size": 63488 00:15:24.004 }, 00:15:24.004 { 00:15:24.004 "name": "BaseBdev2", 00:15:24.004 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:24.004 "is_configured": true, 00:15:24.004 "data_offset": 2048, 00:15:24.004 "data_size": 63488 00:15:24.004 }, 00:15:24.004 { 00:15:24.004 "name": "BaseBdev3", 00:15:24.004 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:24.004 "is_configured": true, 00:15:24.004 "data_offset": 2048, 00:15:24.004 "data_size": 63488 00:15:24.004 } 00:15:24.004 ] 00:15:24.004 }' 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:24.004 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:24.004 [2024-11-20 16:03:22.217691] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:24.004 [2024-11-20 16:03:22.230917] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:24.004 [2024-11-20 16:03:22.230974] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:24.004 [2024-11-20 16:03:22.230992] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:24.004 [2024-11-20 16:03:22.231003] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:24.264 "name": "raid_bdev1", 00:15:24.264 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:24.264 "strip_size_kb": 64, 00:15:24.264 "state": "online", 00:15:24.264 "raid_level": "raid5f", 00:15:24.264 "superblock": true, 00:15:24.264 "num_base_bdevs": 3, 00:15:24.264 "num_base_bdevs_discovered": 2, 00:15:24.264 "num_base_bdevs_operational": 2, 00:15:24.264 "base_bdevs_list": [ 00:15:24.264 { 00:15:24.264 "name": null, 00:15:24.264 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:24.264 "is_configured": false, 00:15:24.264 "data_offset": 0, 00:15:24.264 "data_size": 63488 00:15:24.264 }, 00:15:24.264 { 00:15:24.264 "name": "BaseBdev2", 00:15:24.264 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:24.264 "is_configured": true, 00:15:24.264 "data_offset": 2048, 00:15:24.264 "data_size": 63488 00:15:24.264 }, 00:15:24.264 { 00:15:24.264 "name": "BaseBdev3", 00:15:24.264 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:24.264 "is_configured": true, 00:15:24.264 "data_offset": 2048, 00:15:24.264 "data_size": 63488 00:15:24.264 } 00:15:24.264 ] 00:15:24.264 }' 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:24.264 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:24.525 "name": "raid_bdev1", 00:15:24.525 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:24.525 "strip_size_kb": 64, 00:15:24.525 "state": "online", 00:15:24.525 "raid_level": "raid5f", 00:15:24.525 "superblock": true, 00:15:24.525 "num_base_bdevs": 3, 00:15:24.525 "num_base_bdevs_discovered": 2, 00:15:24.525 "num_base_bdevs_operational": 2, 00:15:24.525 "base_bdevs_list": [ 00:15:24.525 { 00:15:24.525 "name": null, 00:15:24.525 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:24.525 "is_configured": false, 00:15:24.525 "data_offset": 0, 00:15:24.525 "data_size": 63488 00:15:24.525 }, 00:15:24.525 { 00:15:24.525 "name": "BaseBdev2", 00:15:24.525 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:24.525 "is_configured": true, 00:15:24.525 "data_offset": 2048, 00:15:24.525 "data_size": 63488 00:15:24.525 }, 00:15:24.525 { 00:15:24.525 "name": "BaseBdev3", 00:15:24.525 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:24.525 "is_configured": true, 00:15:24.525 "data_offset": 2048, 00:15:24.525 "data_size": 63488 00:15:24.525 } 00:15:24.525 ] 00:15:24.525 }' 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:24.525 [2024-11-20 16:03:22.665776] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:24.525 [2024-11-20 16:03:22.675991] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000029050 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:24.525 16:03:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:24.525 [2024-11-20 16:03:22.681323] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:25.465 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:25.465 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:25.465 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:25.465 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:25.465 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:25.465 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.465 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.465 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:25.465 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:25.465 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:25.465 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:25.465 "name": "raid_bdev1", 00:15:25.465 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:25.465 "strip_size_kb": 64, 00:15:25.465 "state": "online", 00:15:25.465 "raid_level": "raid5f", 00:15:25.465 "superblock": true, 00:15:25.465 "num_base_bdevs": 3, 00:15:25.465 "num_base_bdevs_discovered": 3, 00:15:25.465 "num_base_bdevs_operational": 3, 00:15:25.465 "process": { 00:15:25.465 "type": "rebuild", 00:15:25.465 "target": "spare", 00:15:25.465 "progress": { 00:15:25.465 "blocks": 18432, 00:15:25.465 "percent": 14 00:15:25.465 } 00:15:25.465 }, 00:15:25.465 "base_bdevs_list": [ 00:15:25.465 { 00:15:25.465 "name": "spare", 00:15:25.465 "uuid": "c3caf461-29a0-503c-b7d4-35b7fd697eec", 00:15:25.465 "is_configured": true, 00:15:25.465 "data_offset": 2048, 00:15:25.465 "data_size": 63488 00:15:25.465 }, 00:15:25.465 { 00:15:25.465 "name": "BaseBdev2", 00:15:25.465 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:25.465 "is_configured": true, 00:15:25.465 "data_offset": 2048, 00:15:25.465 "data_size": 63488 00:15:25.465 }, 00:15:25.465 { 00:15:25.465 "name": "BaseBdev3", 00:15:25.465 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:25.465 "is_configured": true, 00:15:25.465 "data_offset": 2048, 00:15:25.465 "data_size": 63488 00:15:25.465 } 00:15:25.465 ] 00:15:25.465 }' 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:15:25.727 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=457 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:25.727 "name": "raid_bdev1", 00:15:25.727 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:25.727 "strip_size_kb": 64, 00:15:25.727 "state": "online", 00:15:25.727 "raid_level": "raid5f", 00:15:25.727 "superblock": true, 00:15:25.727 "num_base_bdevs": 3, 00:15:25.727 "num_base_bdevs_discovered": 3, 00:15:25.727 "num_base_bdevs_operational": 3, 00:15:25.727 "process": { 00:15:25.727 "type": "rebuild", 00:15:25.727 "target": "spare", 00:15:25.727 "progress": { 00:15:25.727 "blocks": 20480, 00:15:25.727 "percent": 16 00:15:25.727 } 00:15:25.727 }, 00:15:25.727 "base_bdevs_list": [ 00:15:25.727 { 00:15:25.727 "name": "spare", 00:15:25.727 "uuid": "c3caf461-29a0-503c-b7d4-35b7fd697eec", 00:15:25.727 "is_configured": true, 00:15:25.727 "data_offset": 2048, 00:15:25.727 "data_size": 63488 00:15:25.727 }, 00:15:25.727 { 00:15:25.727 "name": "BaseBdev2", 00:15:25.727 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:25.727 "is_configured": true, 00:15:25.727 "data_offset": 2048, 00:15:25.727 "data_size": 63488 00:15:25.727 }, 00:15:25.727 { 00:15:25.727 "name": "BaseBdev3", 00:15:25.727 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:25.727 "is_configured": true, 00:15:25.727 "data_offset": 2048, 00:15:25.727 "data_size": 63488 00:15:25.727 } 00:15:25.727 ] 00:15:25.727 }' 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:25.727 16:03:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:26.669 16:03:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:26.669 16:03:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:26.669 16:03:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:26.669 16:03:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:26.669 16:03:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:26.669 16:03:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:26.669 16:03:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:26.669 16:03:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:26.669 16:03:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:26.669 16:03:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:26.669 16:03:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:26.669 16:03:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:26.669 "name": "raid_bdev1", 00:15:26.669 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:26.669 "strip_size_kb": 64, 00:15:26.669 "state": "online", 00:15:26.669 "raid_level": "raid5f", 00:15:26.669 "superblock": true, 00:15:26.669 "num_base_bdevs": 3, 00:15:26.669 "num_base_bdevs_discovered": 3, 00:15:26.669 "num_base_bdevs_operational": 3, 00:15:26.669 "process": { 00:15:26.669 "type": "rebuild", 00:15:26.669 "target": "spare", 00:15:26.669 "progress": { 00:15:26.669 "blocks": 43008, 00:15:26.669 "percent": 33 00:15:26.669 } 00:15:26.669 }, 00:15:26.669 "base_bdevs_list": [ 00:15:26.669 { 00:15:26.669 "name": "spare", 00:15:26.669 "uuid": "c3caf461-29a0-503c-b7d4-35b7fd697eec", 00:15:26.669 "is_configured": true, 00:15:26.669 "data_offset": 2048, 00:15:26.669 "data_size": 63488 00:15:26.669 }, 00:15:26.669 { 00:15:26.669 "name": "BaseBdev2", 00:15:26.669 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:26.669 "is_configured": true, 00:15:26.669 "data_offset": 2048, 00:15:26.669 "data_size": 63488 00:15:26.669 }, 00:15:26.669 { 00:15:26.669 "name": "BaseBdev3", 00:15:26.669 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:26.669 "is_configured": true, 00:15:26.669 "data_offset": 2048, 00:15:26.669 "data_size": 63488 00:15:26.669 } 00:15:26.669 ] 00:15:26.669 }' 00:15:26.669 16:03:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:26.929 16:03:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:26.929 16:03:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:26.929 16:03:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:26.929 16:03:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:27.871 16:03:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:27.871 16:03:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:27.871 16:03:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:27.871 16:03:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:27.871 16:03:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:27.871 16:03:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:27.871 16:03:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.871 16:03:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:27.871 16:03:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.871 16:03:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:27.871 16:03:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.871 16:03:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:27.871 "name": "raid_bdev1", 00:15:27.871 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:27.871 "strip_size_kb": 64, 00:15:27.871 "state": "online", 00:15:27.871 "raid_level": "raid5f", 00:15:27.871 "superblock": true, 00:15:27.871 "num_base_bdevs": 3, 00:15:27.871 "num_base_bdevs_discovered": 3, 00:15:27.871 "num_base_bdevs_operational": 3, 00:15:27.871 "process": { 00:15:27.871 "type": "rebuild", 00:15:27.871 "target": "spare", 00:15:27.871 "progress": { 00:15:27.871 "blocks": 65536, 00:15:27.871 "percent": 51 00:15:27.871 } 00:15:27.871 }, 00:15:27.871 "base_bdevs_list": [ 00:15:27.871 { 00:15:27.871 "name": "spare", 00:15:27.871 "uuid": "c3caf461-29a0-503c-b7d4-35b7fd697eec", 00:15:27.871 "is_configured": true, 00:15:27.871 "data_offset": 2048, 00:15:27.871 "data_size": 63488 00:15:27.871 }, 00:15:27.871 { 00:15:27.871 "name": "BaseBdev2", 00:15:27.871 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:27.871 "is_configured": true, 00:15:27.871 "data_offset": 2048, 00:15:27.871 "data_size": 63488 00:15:27.871 }, 00:15:27.871 { 00:15:27.871 "name": "BaseBdev3", 00:15:27.871 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:27.871 "is_configured": true, 00:15:27.871 "data_offset": 2048, 00:15:27.871 "data_size": 63488 00:15:27.871 } 00:15:27.871 ] 00:15:27.871 }' 00:15:27.871 16:03:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:27.871 16:03:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:27.871 16:03:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:27.871 16:03:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:27.871 16:03:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:28.834 16:03:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:28.834 16:03:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:28.834 16:03:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:28.834 16:03:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:28.834 16:03:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:28.834 16:03:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:28.834 16:03:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:28.834 16:03:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:28.834 16:03:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:28.834 16:03:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:29.094 16:03:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:29.094 16:03:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:29.094 "name": "raid_bdev1", 00:15:29.094 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:29.094 "strip_size_kb": 64, 00:15:29.094 "state": "online", 00:15:29.094 "raid_level": "raid5f", 00:15:29.094 "superblock": true, 00:15:29.094 "num_base_bdevs": 3, 00:15:29.094 "num_base_bdevs_discovered": 3, 00:15:29.094 "num_base_bdevs_operational": 3, 00:15:29.094 "process": { 00:15:29.094 "type": "rebuild", 00:15:29.094 "target": "spare", 00:15:29.094 "progress": { 00:15:29.094 "blocks": 88064, 00:15:29.094 "percent": 69 00:15:29.094 } 00:15:29.094 }, 00:15:29.094 "base_bdevs_list": [ 00:15:29.094 { 00:15:29.094 "name": "spare", 00:15:29.094 "uuid": "c3caf461-29a0-503c-b7d4-35b7fd697eec", 00:15:29.094 "is_configured": true, 00:15:29.094 "data_offset": 2048, 00:15:29.094 "data_size": 63488 00:15:29.094 }, 00:15:29.094 { 00:15:29.094 "name": "BaseBdev2", 00:15:29.094 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:29.094 "is_configured": true, 00:15:29.094 "data_offset": 2048, 00:15:29.094 "data_size": 63488 00:15:29.094 }, 00:15:29.094 { 00:15:29.094 "name": "BaseBdev3", 00:15:29.094 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:29.094 "is_configured": true, 00:15:29.094 "data_offset": 2048, 00:15:29.094 "data_size": 63488 00:15:29.094 } 00:15:29.094 ] 00:15:29.094 }' 00:15:29.094 16:03:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:29.094 16:03:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:29.094 16:03:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:29.094 16:03:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:29.094 16:03:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:30.036 16:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:30.036 16:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:30.036 16:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:30.036 16:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:30.036 16:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:30.036 16:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:30.036 16:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:30.036 16:03:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:30.036 16:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:30.036 16:03:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:30.036 16:03:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:30.036 16:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:30.036 "name": "raid_bdev1", 00:15:30.036 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:30.036 "strip_size_kb": 64, 00:15:30.036 "state": "online", 00:15:30.036 "raid_level": "raid5f", 00:15:30.036 "superblock": true, 00:15:30.036 "num_base_bdevs": 3, 00:15:30.036 "num_base_bdevs_discovered": 3, 00:15:30.036 "num_base_bdevs_operational": 3, 00:15:30.036 "process": { 00:15:30.036 "type": "rebuild", 00:15:30.036 "target": "spare", 00:15:30.036 "progress": { 00:15:30.036 "blocks": 110592, 00:15:30.036 "percent": 87 00:15:30.036 } 00:15:30.036 }, 00:15:30.036 "base_bdevs_list": [ 00:15:30.036 { 00:15:30.036 "name": "spare", 00:15:30.036 "uuid": "c3caf461-29a0-503c-b7d4-35b7fd697eec", 00:15:30.036 "is_configured": true, 00:15:30.036 "data_offset": 2048, 00:15:30.036 "data_size": 63488 00:15:30.036 }, 00:15:30.036 { 00:15:30.036 "name": "BaseBdev2", 00:15:30.036 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:30.036 "is_configured": true, 00:15:30.036 "data_offset": 2048, 00:15:30.036 "data_size": 63488 00:15:30.036 }, 00:15:30.036 { 00:15:30.036 "name": "BaseBdev3", 00:15:30.036 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:30.036 "is_configured": true, 00:15:30.036 "data_offset": 2048, 00:15:30.036 "data_size": 63488 00:15:30.036 } 00:15:30.036 ] 00:15:30.036 }' 00:15:30.036 16:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:30.036 16:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:30.036 16:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:30.036 16:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:30.036 16:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:30.979 [2024-11-20 16:03:28.938158] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:30.979 [2024-11-20 16:03:28.938249] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:30.979 [2024-11-20 16:03:28.938376] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:31.241 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:31.241 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:31.241 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:31.241 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:31.241 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:31.241 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:31.241 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:31.241 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:31.241 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.241 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.241 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.241 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:31.241 "name": "raid_bdev1", 00:15:31.241 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:31.241 "strip_size_kb": 64, 00:15:31.241 "state": "online", 00:15:31.241 "raid_level": "raid5f", 00:15:31.241 "superblock": true, 00:15:31.241 "num_base_bdevs": 3, 00:15:31.241 "num_base_bdevs_discovered": 3, 00:15:31.241 "num_base_bdevs_operational": 3, 00:15:31.241 "base_bdevs_list": [ 00:15:31.241 { 00:15:31.241 "name": "spare", 00:15:31.241 "uuid": "c3caf461-29a0-503c-b7d4-35b7fd697eec", 00:15:31.241 "is_configured": true, 00:15:31.241 "data_offset": 2048, 00:15:31.241 "data_size": 63488 00:15:31.241 }, 00:15:31.241 { 00:15:31.241 "name": "BaseBdev2", 00:15:31.241 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:31.241 "is_configured": true, 00:15:31.241 "data_offset": 2048, 00:15:31.241 "data_size": 63488 00:15:31.241 }, 00:15:31.241 { 00:15:31.242 "name": "BaseBdev3", 00:15:31.242 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:31.242 "is_configured": true, 00:15:31.242 "data_offset": 2048, 00:15:31.242 "data_size": 63488 00:15:31.242 } 00:15:31.242 ] 00:15:31.242 }' 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:31.242 "name": "raid_bdev1", 00:15:31.242 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:31.242 "strip_size_kb": 64, 00:15:31.242 "state": "online", 00:15:31.242 "raid_level": "raid5f", 00:15:31.242 "superblock": true, 00:15:31.242 "num_base_bdevs": 3, 00:15:31.242 "num_base_bdevs_discovered": 3, 00:15:31.242 "num_base_bdevs_operational": 3, 00:15:31.242 "base_bdevs_list": [ 00:15:31.242 { 00:15:31.242 "name": "spare", 00:15:31.242 "uuid": "c3caf461-29a0-503c-b7d4-35b7fd697eec", 00:15:31.242 "is_configured": true, 00:15:31.242 "data_offset": 2048, 00:15:31.242 "data_size": 63488 00:15:31.242 }, 00:15:31.242 { 00:15:31.242 "name": "BaseBdev2", 00:15:31.242 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:31.242 "is_configured": true, 00:15:31.242 "data_offset": 2048, 00:15:31.242 "data_size": 63488 00:15:31.242 }, 00:15:31.242 { 00:15:31.242 "name": "BaseBdev3", 00:15:31.242 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:31.242 "is_configured": true, 00:15:31.242 "data_offset": 2048, 00:15:31.242 "data_size": 63488 00:15:31.242 } 00:15:31.242 ] 00:15:31.242 }' 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:31.242 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.501 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:31.501 "name": "raid_bdev1", 00:15:31.501 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:31.501 "strip_size_kb": 64, 00:15:31.501 "state": "online", 00:15:31.501 "raid_level": "raid5f", 00:15:31.501 "superblock": true, 00:15:31.501 "num_base_bdevs": 3, 00:15:31.501 "num_base_bdevs_discovered": 3, 00:15:31.501 "num_base_bdevs_operational": 3, 00:15:31.501 "base_bdevs_list": [ 00:15:31.501 { 00:15:31.501 "name": "spare", 00:15:31.501 "uuid": "c3caf461-29a0-503c-b7d4-35b7fd697eec", 00:15:31.501 "is_configured": true, 00:15:31.501 "data_offset": 2048, 00:15:31.501 "data_size": 63488 00:15:31.501 }, 00:15:31.501 { 00:15:31.501 "name": "BaseBdev2", 00:15:31.501 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:31.501 "is_configured": true, 00:15:31.501 "data_offset": 2048, 00:15:31.501 "data_size": 63488 00:15:31.501 }, 00:15:31.501 { 00:15:31.501 "name": "BaseBdev3", 00:15:31.501 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:31.501 "is_configured": true, 00:15:31.501 "data_offset": 2048, 00:15:31.501 "data_size": 63488 00:15:31.501 } 00:15:31.501 ] 00:15:31.501 }' 00:15:31.501 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:31.501 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.762 [2024-11-20 16:03:29.805047] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:31.762 [2024-11-20 16:03:29.805185] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:31.762 [2024-11-20 16:03:29.805277] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:31.762 [2024-11-20 16:03:29.805357] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:31.762 [2024-11-20 16:03:29.805379] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:31.762 16:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:32.064 /dev/nbd0 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:32.064 1+0 records in 00:15:32.064 1+0 records out 00:15:32.064 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000345479 s, 11.9 MB/s 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:32.064 /dev/nbd1 00:15:32.064 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:32.326 1+0 records in 00:15:32.326 1+0 records out 00:15:32.326 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000457748 s, 8.9 MB/s 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:32.326 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:32.587 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:32.587 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:32.587 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:32.587 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:32.587 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:32.587 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:32.587 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:32.587 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:32.587 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:32.587 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:32.848 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:32.848 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:32.848 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:32.849 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:32.849 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:32.849 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:32.849 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:32.849 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:32.849 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:15:32.849 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:15:32.849 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.849 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:32.849 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:32.849 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:32.849 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.849 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:32.849 [2024-11-20 16:03:30.919931] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:32.849 [2024-11-20 16:03:30.919988] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:32.849 [2024-11-20 16:03:30.920009] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:15:32.849 [2024-11-20 16:03:30.920020] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:32.849 [2024-11-20 16:03:30.922277] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:32.849 [2024-11-20 16:03:30.922314] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:32.849 [2024-11-20 16:03:30.922401] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:32.849 [2024-11-20 16:03:30.922447] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:32.849 [2024-11-20 16:03:30.922569] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:32.849 [2024-11-20 16:03:30.922678] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:32.849 spare 00:15:32.849 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:32.849 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:15:32.849 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.849 16:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:32.849 [2024-11-20 16:03:31.022769] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:15:32.849 [2024-11-20 16:03:31.022981] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:15:32.849 [2024-11-20 16:03:31.023307] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000047700 00:15:32.849 [2024-11-20 16:03:31.026828] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:15:32.849 [2024-11-20 16:03:31.026846] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:15:32.849 [2024-11-20 16:03:31.027036] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:32.849 "name": "raid_bdev1", 00:15:32.849 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:32.849 "strip_size_kb": 64, 00:15:32.849 "state": "online", 00:15:32.849 "raid_level": "raid5f", 00:15:32.849 "superblock": true, 00:15:32.849 "num_base_bdevs": 3, 00:15:32.849 "num_base_bdevs_discovered": 3, 00:15:32.849 "num_base_bdevs_operational": 3, 00:15:32.849 "base_bdevs_list": [ 00:15:32.849 { 00:15:32.849 "name": "spare", 00:15:32.849 "uuid": "c3caf461-29a0-503c-b7d4-35b7fd697eec", 00:15:32.849 "is_configured": true, 00:15:32.849 "data_offset": 2048, 00:15:32.849 "data_size": 63488 00:15:32.849 }, 00:15:32.849 { 00:15:32.849 "name": "BaseBdev2", 00:15:32.849 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:32.849 "is_configured": true, 00:15:32.849 "data_offset": 2048, 00:15:32.849 "data_size": 63488 00:15:32.849 }, 00:15:32.849 { 00:15:32.849 "name": "BaseBdev3", 00:15:32.849 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:32.849 "is_configured": true, 00:15:32.849 "data_offset": 2048, 00:15:32.849 "data_size": 63488 00:15:32.849 } 00:15:32.849 ] 00:15:32.849 }' 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:32.849 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.112 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:33.112 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:33.112 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:33.112 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:33.112 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:33.112 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:33.112 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:33.112 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.112 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:33.374 "name": "raid_bdev1", 00:15:33.374 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:33.374 "strip_size_kb": 64, 00:15:33.374 "state": "online", 00:15:33.374 "raid_level": "raid5f", 00:15:33.374 "superblock": true, 00:15:33.374 "num_base_bdevs": 3, 00:15:33.374 "num_base_bdevs_discovered": 3, 00:15:33.374 "num_base_bdevs_operational": 3, 00:15:33.374 "base_bdevs_list": [ 00:15:33.374 { 00:15:33.374 "name": "spare", 00:15:33.374 "uuid": "c3caf461-29a0-503c-b7d4-35b7fd697eec", 00:15:33.374 "is_configured": true, 00:15:33.374 "data_offset": 2048, 00:15:33.374 "data_size": 63488 00:15:33.374 }, 00:15:33.374 { 00:15:33.374 "name": "BaseBdev2", 00:15:33.374 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:33.374 "is_configured": true, 00:15:33.374 "data_offset": 2048, 00:15:33.374 "data_size": 63488 00:15:33.374 }, 00:15:33.374 { 00:15:33.374 "name": "BaseBdev3", 00:15:33.374 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:33.374 "is_configured": true, 00:15:33.374 "data_offset": 2048, 00:15:33.374 "data_size": 63488 00:15:33.374 } 00:15:33.374 ] 00:15:33.374 }' 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.374 [2024-11-20 16:03:31.483184] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:33.374 "name": "raid_bdev1", 00:15:33.374 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:33.374 "strip_size_kb": 64, 00:15:33.374 "state": "online", 00:15:33.374 "raid_level": "raid5f", 00:15:33.374 "superblock": true, 00:15:33.374 "num_base_bdevs": 3, 00:15:33.374 "num_base_bdevs_discovered": 2, 00:15:33.374 "num_base_bdevs_operational": 2, 00:15:33.374 "base_bdevs_list": [ 00:15:33.374 { 00:15:33.374 "name": null, 00:15:33.374 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:33.374 "is_configured": false, 00:15:33.374 "data_offset": 0, 00:15:33.374 "data_size": 63488 00:15:33.374 }, 00:15:33.374 { 00:15:33.374 "name": "BaseBdev2", 00:15:33.374 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:33.374 "is_configured": true, 00:15:33.374 "data_offset": 2048, 00:15:33.374 "data_size": 63488 00:15:33.374 }, 00:15:33.374 { 00:15:33.374 "name": "BaseBdev3", 00:15:33.374 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:33.374 "is_configured": true, 00:15:33.374 "data_offset": 2048, 00:15:33.374 "data_size": 63488 00:15:33.374 } 00:15:33.374 ] 00:15:33.374 }' 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:33.374 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.664 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:33.664 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.664 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.664 [2024-11-20 16:03:31.787303] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:33.664 [2024-11-20 16:03:31.787481] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:33.664 [2024-11-20 16:03:31.787499] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:33.664 [2024-11-20 16:03:31.787537] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:33.664 [2024-11-20 16:03:31.797464] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000477d0 00:15:33.664 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.664 16:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:15:33.664 [2024-11-20 16:03:31.802895] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:34.605 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:34.606 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:34.606 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:34.606 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:34.606 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:34.606 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:34.606 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:34.606 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:34.606 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:34.606 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.606 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:34.606 "name": "raid_bdev1", 00:15:34.606 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:34.606 "strip_size_kb": 64, 00:15:34.606 "state": "online", 00:15:34.606 "raid_level": "raid5f", 00:15:34.606 "superblock": true, 00:15:34.606 "num_base_bdevs": 3, 00:15:34.606 "num_base_bdevs_discovered": 3, 00:15:34.606 "num_base_bdevs_operational": 3, 00:15:34.606 "process": { 00:15:34.606 "type": "rebuild", 00:15:34.606 "target": "spare", 00:15:34.606 "progress": { 00:15:34.606 "blocks": 18432, 00:15:34.606 "percent": 14 00:15:34.606 } 00:15:34.606 }, 00:15:34.606 "base_bdevs_list": [ 00:15:34.606 { 00:15:34.606 "name": "spare", 00:15:34.606 "uuid": "c3caf461-29a0-503c-b7d4-35b7fd697eec", 00:15:34.606 "is_configured": true, 00:15:34.606 "data_offset": 2048, 00:15:34.606 "data_size": 63488 00:15:34.606 }, 00:15:34.606 { 00:15:34.606 "name": "BaseBdev2", 00:15:34.606 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:34.606 "is_configured": true, 00:15:34.606 "data_offset": 2048, 00:15:34.606 "data_size": 63488 00:15:34.606 }, 00:15:34.606 { 00:15:34.606 "name": "BaseBdev3", 00:15:34.606 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:34.606 "is_configured": true, 00:15:34.606 "data_offset": 2048, 00:15:34.606 "data_size": 63488 00:15:34.606 } 00:15:34.606 ] 00:15:34.606 }' 00:15:34.606 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:34.901 [2024-11-20 16:03:32.904180] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:34.901 [2024-11-20 16:03:32.913326] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:34.901 [2024-11-20 16:03:32.913386] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:34.901 [2024-11-20 16:03:32.913402] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:34.901 [2024-11-20 16:03:32.913412] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:34.901 "name": "raid_bdev1", 00:15:34.901 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:34.901 "strip_size_kb": 64, 00:15:34.901 "state": "online", 00:15:34.901 "raid_level": "raid5f", 00:15:34.901 "superblock": true, 00:15:34.901 "num_base_bdevs": 3, 00:15:34.901 "num_base_bdevs_discovered": 2, 00:15:34.901 "num_base_bdevs_operational": 2, 00:15:34.901 "base_bdevs_list": [ 00:15:34.901 { 00:15:34.901 "name": null, 00:15:34.901 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:34.901 "is_configured": false, 00:15:34.901 "data_offset": 0, 00:15:34.901 "data_size": 63488 00:15:34.901 }, 00:15:34.901 { 00:15:34.901 "name": "BaseBdev2", 00:15:34.901 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:34.901 "is_configured": true, 00:15:34.901 "data_offset": 2048, 00:15:34.901 "data_size": 63488 00:15:34.901 }, 00:15:34.901 { 00:15:34.901 "name": "BaseBdev3", 00:15:34.901 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:34.901 "is_configured": true, 00:15:34.901 "data_offset": 2048, 00:15:34.901 "data_size": 63488 00:15:34.901 } 00:15:34.901 ] 00:15:34.901 }' 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:34.901 16:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.161 16:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:35.161 16:03:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:35.161 16:03:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.161 [2024-11-20 16:03:33.263643] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:35.161 [2024-11-20 16:03:33.263712] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:35.161 [2024-11-20 16:03:33.263731] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:15:35.161 [2024-11-20 16:03:33.263743] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:35.161 [2024-11-20 16:03:33.264190] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:35.161 [2024-11-20 16:03:33.264207] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:35.161 [2024-11-20 16:03:33.264294] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:35.161 [2024-11-20 16:03:33.264322] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:35.161 [2024-11-20 16:03:33.264332] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:35.161 [2024-11-20 16:03:33.264351] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:35.161 [2024-11-20 16:03:33.274464] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000478a0 00:15:35.161 spare 00:15:35.161 16:03:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:35.161 16:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:15:35.161 [2024-11-20 16:03:33.279779] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:36.102 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:36.102 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:36.102 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:36.102 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:36.102 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:36.102 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:36.102 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.102 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.102 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:36.102 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.102 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:36.102 "name": "raid_bdev1", 00:15:36.102 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:36.102 "strip_size_kb": 64, 00:15:36.102 "state": "online", 00:15:36.102 "raid_level": "raid5f", 00:15:36.102 "superblock": true, 00:15:36.102 "num_base_bdevs": 3, 00:15:36.102 "num_base_bdevs_discovered": 3, 00:15:36.102 "num_base_bdevs_operational": 3, 00:15:36.102 "process": { 00:15:36.102 "type": "rebuild", 00:15:36.102 "target": "spare", 00:15:36.102 "progress": { 00:15:36.102 "blocks": 18432, 00:15:36.102 "percent": 14 00:15:36.102 } 00:15:36.102 }, 00:15:36.102 "base_bdevs_list": [ 00:15:36.102 { 00:15:36.102 "name": "spare", 00:15:36.102 "uuid": "c3caf461-29a0-503c-b7d4-35b7fd697eec", 00:15:36.102 "is_configured": true, 00:15:36.102 "data_offset": 2048, 00:15:36.102 "data_size": 63488 00:15:36.102 }, 00:15:36.102 { 00:15:36.102 "name": "BaseBdev2", 00:15:36.102 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:36.102 "is_configured": true, 00:15:36.102 "data_offset": 2048, 00:15:36.102 "data_size": 63488 00:15:36.102 }, 00:15:36.102 { 00:15:36.102 "name": "BaseBdev3", 00:15:36.102 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:36.102 "is_configured": true, 00:15:36.102 "data_offset": 2048, 00:15:36.102 "data_size": 63488 00:15:36.102 } 00:15:36.102 ] 00:15:36.102 }' 00:15:36.102 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:36.102 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:36.102 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:36.362 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:36.362 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:15:36.362 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.362 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:36.362 [2024-11-20 16:03:34.384737] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:36.362 [2024-11-20 16:03:34.389856] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:36.362 [2024-11-20 16:03:34.390057] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:36.362 [2024-11-20 16:03:34.390083] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:36.363 [2024-11-20 16:03:34.390093] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:36.363 "name": "raid_bdev1", 00:15:36.363 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:36.363 "strip_size_kb": 64, 00:15:36.363 "state": "online", 00:15:36.363 "raid_level": "raid5f", 00:15:36.363 "superblock": true, 00:15:36.363 "num_base_bdevs": 3, 00:15:36.363 "num_base_bdevs_discovered": 2, 00:15:36.363 "num_base_bdevs_operational": 2, 00:15:36.363 "base_bdevs_list": [ 00:15:36.363 { 00:15:36.363 "name": null, 00:15:36.363 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:36.363 "is_configured": false, 00:15:36.363 "data_offset": 0, 00:15:36.363 "data_size": 63488 00:15:36.363 }, 00:15:36.363 { 00:15:36.363 "name": "BaseBdev2", 00:15:36.363 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:36.363 "is_configured": true, 00:15:36.363 "data_offset": 2048, 00:15:36.363 "data_size": 63488 00:15:36.363 }, 00:15:36.363 { 00:15:36.363 "name": "BaseBdev3", 00:15:36.363 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:36.363 "is_configured": true, 00:15:36.363 "data_offset": 2048, 00:15:36.363 "data_size": 63488 00:15:36.363 } 00:15:36.363 ] 00:15:36.363 }' 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:36.363 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:36.624 "name": "raid_bdev1", 00:15:36.624 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:36.624 "strip_size_kb": 64, 00:15:36.624 "state": "online", 00:15:36.624 "raid_level": "raid5f", 00:15:36.624 "superblock": true, 00:15:36.624 "num_base_bdevs": 3, 00:15:36.624 "num_base_bdevs_discovered": 2, 00:15:36.624 "num_base_bdevs_operational": 2, 00:15:36.624 "base_bdevs_list": [ 00:15:36.624 { 00:15:36.624 "name": null, 00:15:36.624 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:36.624 "is_configured": false, 00:15:36.624 "data_offset": 0, 00:15:36.624 "data_size": 63488 00:15:36.624 }, 00:15:36.624 { 00:15:36.624 "name": "BaseBdev2", 00:15:36.624 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:36.624 "is_configured": true, 00:15:36.624 "data_offset": 2048, 00:15:36.624 "data_size": 63488 00:15:36.624 }, 00:15:36.624 { 00:15:36.624 "name": "BaseBdev3", 00:15:36.624 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:36.624 "is_configured": true, 00:15:36.624 "data_offset": 2048, 00:15:36.624 "data_size": 63488 00:15:36.624 } 00:15:36.624 ] 00:15:36.624 }' 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:36.624 [2024-11-20 16:03:34.828357] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:36.624 [2024-11-20 16:03:34.828409] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:36.624 [2024-11-20 16:03:34.828431] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:15:36.624 [2024-11-20 16:03:34.828441] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:36.624 [2024-11-20 16:03:34.828884] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:36.624 [2024-11-20 16:03:34.828904] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:36.624 [2024-11-20 16:03:34.828982] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:15:36.624 [2024-11-20 16:03:34.828999] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:36.624 [2024-11-20 16:03:34.829008] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:36.624 [2024-11-20 16:03:34.829030] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:15:36.624 BaseBdev1 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.624 16:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:38.010 "name": "raid_bdev1", 00:15:38.010 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:38.010 "strip_size_kb": 64, 00:15:38.010 "state": "online", 00:15:38.010 "raid_level": "raid5f", 00:15:38.010 "superblock": true, 00:15:38.010 "num_base_bdevs": 3, 00:15:38.010 "num_base_bdevs_discovered": 2, 00:15:38.010 "num_base_bdevs_operational": 2, 00:15:38.010 "base_bdevs_list": [ 00:15:38.010 { 00:15:38.010 "name": null, 00:15:38.010 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.010 "is_configured": false, 00:15:38.010 "data_offset": 0, 00:15:38.010 "data_size": 63488 00:15:38.010 }, 00:15:38.010 { 00:15:38.010 "name": "BaseBdev2", 00:15:38.010 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:38.010 "is_configured": true, 00:15:38.010 "data_offset": 2048, 00:15:38.010 "data_size": 63488 00:15:38.010 }, 00:15:38.010 { 00:15:38.010 "name": "BaseBdev3", 00:15:38.010 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:38.010 "is_configured": true, 00:15:38.010 "data_offset": 2048, 00:15:38.010 "data_size": 63488 00:15:38.010 } 00:15:38.010 ] 00:15:38.010 }' 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:38.010 16:03:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:38.010 "name": "raid_bdev1", 00:15:38.010 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:38.010 "strip_size_kb": 64, 00:15:38.010 "state": "online", 00:15:38.010 "raid_level": "raid5f", 00:15:38.010 "superblock": true, 00:15:38.010 "num_base_bdevs": 3, 00:15:38.010 "num_base_bdevs_discovered": 2, 00:15:38.010 "num_base_bdevs_operational": 2, 00:15:38.010 "base_bdevs_list": [ 00:15:38.010 { 00:15:38.010 "name": null, 00:15:38.010 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.010 "is_configured": false, 00:15:38.010 "data_offset": 0, 00:15:38.010 "data_size": 63488 00:15:38.010 }, 00:15:38.010 { 00:15:38.010 "name": "BaseBdev2", 00:15:38.010 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:38.010 "is_configured": true, 00:15:38.010 "data_offset": 2048, 00:15:38.010 "data_size": 63488 00:15:38.010 }, 00:15:38.010 { 00:15:38.010 "name": "BaseBdev3", 00:15:38.010 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:38.010 "is_configured": true, 00:15:38.010 "data_offset": 2048, 00:15:38.010 "data_size": 63488 00:15:38.010 } 00:15:38.010 ] 00:15:38.010 }' 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:38.010 [2024-11-20 16:03:36.244745] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:38.010 [2024-11-20 16:03:36.244888] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:38.010 [2024-11-20 16:03:36.244903] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:38.010 request: 00:15:38.010 { 00:15:38.010 "base_bdev": "BaseBdev1", 00:15:38.010 "raid_bdev": "raid_bdev1", 00:15:38.010 "method": "bdev_raid_add_base_bdev", 00:15:38.010 "req_id": 1 00:15:38.010 } 00:15:38.010 Got JSON-RPC error response 00:15:38.010 response: 00:15:38.010 { 00:15:38.010 "code": -22, 00:15:38.010 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:15:38.010 } 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:38.010 16:03:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:39.398 "name": "raid_bdev1", 00:15:39.398 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:39.398 "strip_size_kb": 64, 00:15:39.398 "state": "online", 00:15:39.398 "raid_level": "raid5f", 00:15:39.398 "superblock": true, 00:15:39.398 "num_base_bdevs": 3, 00:15:39.398 "num_base_bdevs_discovered": 2, 00:15:39.398 "num_base_bdevs_operational": 2, 00:15:39.398 "base_bdevs_list": [ 00:15:39.398 { 00:15:39.398 "name": null, 00:15:39.398 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:39.398 "is_configured": false, 00:15:39.398 "data_offset": 0, 00:15:39.398 "data_size": 63488 00:15:39.398 }, 00:15:39.398 { 00:15:39.398 "name": "BaseBdev2", 00:15:39.398 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:39.398 "is_configured": true, 00:15:39.398 "data_offset": 2048, 00:15:39.398 "data_size": 63488 00:15:39.398 }, 00:15:39.398 { 00:15:39.398 "name": "BaseBdev3", 00:15:39.398 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:39.398 "is_configured": true, 00:15:39.398 "data_offset": 2048, 00:15:39.398 "data_size": 63488 00:15:39.398 } 00:15:39.398 ] 00:15:39.398 }' 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:39.398 "name": "raid_bdev1", 00:15:39.398 "uuid": "3c13d56a-e6e8-41cc-bbeb-964af58dab32", 00:15:39.398 "strip_size_kb": 64, 00:15:39.398 "state": "online", 00:15:39.398 "raid_level": "raid5f", 00:15:39.398 "superblock": true, 00:15:39.398 "num_base_bdevs": 3, 00:15:39.398 "num_base_bdevs_discovered": 2, 00:15:39.398 "num_base_bdevs_operational": 2, 00:15:39.398 "base_bdevs_list": [ 00:15:39.398 { 00:15:39.398 "name": null, 00:15:39.398 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:39.398 "is_configured": false, 00:15:39.398 "data_offset": 0, 00:15:39.398 "data_size": 63488 00:15:39.398 }, 00:15:39.398 { 00:15:39.398 "name": "BaseBdev2", 00:15:39.398 "uuid": "01611655-822b-5794-a56b-b80de1058349", 00:15:39.398 "is_configured": true, 00:15:39.398 "data_offset": 2048, 00:15:39.398 "data_size": 63488 00:15:39.398 }, 00:15:39.398 { 00:15:39.398 "name": "BaseBdev3", 00:15:39.398 "uuid": "49885d0e-9b53-5687-8a14-d061e443c5a1", 00:15:39.398 "is_configured": true, 00:15:39.398 "data_offset": 2048, 00:15:39.398 "data_size": 63488 00:15:39.398 } 00:15:39.398 ] 00:15:39.398 }' 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:39.398 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:39.658 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:39.658 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 79759 00:15:39.658 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 79759 ']' 00:15:39.658 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 79759 00:15:39.658 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:15:39.658 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:39.658 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79759 00:15:39.658 killing process with pid 79759 00:15:39.658 Received shutdown signal, test time was about 60.000000 seconds 00:15:39.658 00:15:39.658 Latency(us) 00:15:39.658 [2024-11-20T16:03:37.909Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:39.658 [2024-11-20T16:03:37.909Z] =================================================================================================================== 00:15:39.658 [2024-11-20T16:03:37.909Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:39.658 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:39.658 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:39.658 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79759' 00:15:39.658 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 79759 00:15:39.658 [2024-11-20 16:03:37.690088] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:39.658 16:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 79759 00:15:39.658 [2024-11-20 16:03:37.690204] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:39.658 [2024-11-20 16:03:37.690267] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:39.658 [2024-11-20 16:03:37.690278] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:15:39.918 [2024-11-20 16:03:37.936292] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:40.488 16:03:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:15:40.488 00:15:40.488 real 0m20.416s 00:15:40.488 user 0m25.292s 00:15:40.488 sys 0m2.139s 00:15:40.488 ************************************ 00:15:40.488 END TEST raid5f_rebuild_test_sb 00:15:40.488 ************************************ 00:15:40.488 16:03:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:40.488 16:03:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:40.488 16:03:38 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:15:40.488 16:03:38 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 4 false 00:15:40.488 16:03:38 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:15:40.488 16:03:38 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:40.488 16:03:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:40.488 ************************************ 00:15:40.488 START TEST raid5f_state_function_test 00:15:40.488 ************************************ 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 4 false 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:15:40.488 Process raid pid: 80479 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=80479 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 80479' 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 80479 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 80479 ']' 00:15:40.488 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:40.488 16:03:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:40.748 [2024-11-20 16:03:38.774013] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:15:40.748 [2024-11-20 16:03:38.774137] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:40.748 [2024-11-20 16:03:38.935306] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:41.009 [2024-11-20 16:03:39.037906] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:41.009 [2024-11-20 16:03:39.175837] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:41.009 [2024-11-20 16:03:39.175890] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.581 [2024-11-20 16:03:39.679752] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:41.581 [2024-11-20 16:03:39.679798] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:41.581 [2024-11-20 16:03:39.679808] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:41.581 [2024-11-20 16:03:39.679817] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:41.581 [2024-11-20 16:03:39.679824] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:41.581 [2024-11-20 16:03:39.679832] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:41.581 [2024-11-20 16:03:39.679838] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:15:41.581 [2024-11-20 16:03:39.679847] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:41.581 "name": "Existed_Raid", 00:15:41.581 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:41.581 "strip_size_kb": 64, 00:15:41.581 "state": "configuring", 00:15:41.581 "raid_level": "raid5f", 00:15:41.581 "superblock": false, 00:15:41.581 "num_base_bdevs": 4, 00:15:41.581 "num_base_bdevs_discovered": 0, 00:15:41.581 "num_base_bdevs_operational": 4, 00:15:41.581 "base_bdevs_list": [ 00:15:41.581 { 00:15:41.581 "name": "BaseBdev1", 00:15:41.581 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:41.581 "is_configured": false, 00:15:41.581 "data_offset": 0, 00:15:41.581 "data_size": 0 00:15:41.581 }, 00:15:41.581 { 00:15:41.581 "name": "BaseBdev2", 00:15:41.581 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:41.581 "is_configured": false, 00:15:41.581 "data_offset": 0, 00:15:41.581 "data_size": 0 00:15:41.581 }, 00:15:41.581 { 00:15:41.581 "name": "BaseBdev3", 00:15:41.581 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:41.581 "is_configured": false, 00:15:41.581 "data_offset": 0, 00:15:41.581 "data_size": 0 00:15:41.581 }, 00:15:41.581 { 00:15:41.581 "name": "BaseBdev4", 00:15:41.581 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:41.581 "is_configured": false, 00:15:41.581 "data_offset": 0, 00:15:41.581 "data_size": 0 00:15:41.581 } 00:15:41.581 ] 00:15:41.581 }' 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:41.581 16:03:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.842 [2024-11-20 16:03:40.015769] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:41.842 [2024-11-20 16:03:40.015802] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.842 [2024-11-20 16:03:40.023775] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:41.842 [2024-11-20 16:03:40.023813] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:41.842 [2024-11-20 16:03:40.023822] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:41.842 [2024-11-20 16:03:40.023831] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:41.842 [2024-11-20 16:03:40.023838] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:41.842 [2024-11-20 16:03:40.023846] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:41.842 [2024-11-20 16:03:40.023852] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:15:41.842 [2024-11-20 16:03:40.023861] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.842 [2024-11-20 16:03:40.056324] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:41.842 BaseBdev1 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.842 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.842 [ 00:15:41.842 { 00:15:41.842 "name": "BaseBdev1", 00:15:41.842 "aliases": [ 00:15:41.842 "23a1a968-4da8-4d6c-87a5-c96cdcbf0898" 00:15:41.842 ], 00:15:41.843 "product_name": "Malloc disk", 00:15:41.843 "block_size": 512, 00:15:41.843 "num_blocks": 65536, 00:15:41.843 "uuid": "23a1a968-4da8-4d6c-87a5-c96cdcbf0898", 00:15:41.843 "assigned_rate_limits": { 00:15:41.843 "rw_ios_per_sec": 0, 00:15:41.843 "rw_mbytes_per_sec": 0, 00:15:41.843 "r_mbytes_per_sec": 0, 00:15:41.843 "w_mbytes_per_sec": 0 00:15:41.843 }, 00:15:41.843 "claimed": true, 00:15:41.843 "claim_type": "exclusive_write", 00:15:41.843 "zoned": false, 00:15:41.843 "supported_io_types": { 00:15:41.843 "read": true, 00:15:41.843 "write": true, 00:15:41.843 "unmap": true, 00:15:41.843 "flush": true, 00:15:41.843 "reset": true, 00:15:41.843 "nvme_admin": false, 00:15:41.843 "nvme_io": false, 00:15:41.843 "nvme_io_md": false, 00:15:41.843 "write_zeroes": true, 00:15:41.843 "zcopy": true, 00:15:41.843 "get_zone_info": false, 00:15:41.843 "zone_management": false, 00:15:41.843 "zone_append": false, 00:15:41.843 "compare": false, 00:15:41.843 "compare_and_write": false, 00:15:41.843 "abort": true, 00:15:41.843 "seek_hole": false, 00:15:41.843 "seek_data": false, 00:15:41.843 "copy": true, 00:15:41.843 "nvme_iov_md": false 00:15:41.843 }, 00:15:41.843 "memory_domains": [ 00:15:41.843 { 00:15:41.843 "dma_device_id": "system", 00:15:41.843 "dma_device_type": 1 00:15:41.843 }, 00:15:41.843 { 00:15:41.843 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:41.843 "dma_device_type": 2 00:15:41.843 } 00:15:41.843 ], 00:15:41.843 "driver_specific": {} 00:15:41.843 } 00:15:41.843 ] 00:15:41.843 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.843 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:41.843 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:41.843 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:41.843 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:41.843 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:41.843 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:41.843 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:41.843 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:41.843 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:41.843 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:41.843 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:41.843 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.843 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.843 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.843 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:42.103 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.103 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:42.103 "name": "Existed_Raid", 00:15:42.103 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.103 "strip_size_kb": 64, 00:15:42.103 "state": "configuring", 00:15:42.103 "raid_level": "raid5f", 00:15:42.103 "superblock": false, 00:15:42.103 "num_base_bdevs": 4, 00:15:42.103 "num_base_bdevs_discovered": 1, 00:15:42.103 "num_base_bdevs_operational": 4, 00:15:42.103 "base_bdevs_list": [ 00:15:42.103 { 00:15:42.103 "name": "BaseBdev1", 00:15:42.103 "uuid": "23a1a968-4da8-4d6c-87a5-c96cdcbf0898", 00:15:42.103 "is_configured": true, 00:15:42.103 "data_offset": 0, 00:15:42.103 "data_size": 65536 00:15:42.103 }, 00:15:42.103 { 00:15:42.103 "name": "BaseBdev2", 00:15:42.103 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.103 "is_configured": false, 00:15:42.103 "data_offset": 0, 00:15:42.103 "data_size": 0 00:15:42.103 }, 00:15:42.103 { 00:15:42.103 "name": "BaseBdev3", 00:15:42.103 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.103 "is_configured": false, 00:15:42.103 "data_offset": 0, 00:15:42.103 "data_size": 0 00:15:42.103 }, 00:15:42.103 { 00:15:42.103 "name": "BaseBdev4", 00:15:42.103 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.103 "is_configured": false, 00:15:42.103 "data_offset": 0, 00:15:42.103 "data_size": 0 00:15:42.103 } 00:15:42.103 ] 00:15:42.103 }' 00:15:42.103 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:42.103 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.364 [2024-11-20 16:03:40.400463] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:42.364 [2024-11-20 16:03:40.400511] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.364 [2024-11-20 16:03:40.408508] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:42.364 [2024-11-20 16:03:40.410438] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:42.364 [2024-11-20 16:03:40.410559] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:42.364 [2024-11-20 16:03:40.410617] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:42.364 [2024-11-20 16:03:40.410648] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:42.364 [2024-11-20 16:03:40.410679] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:15:42.364 [2024-11-20 16:03:40.410704] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.364 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.365 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.365 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:42.365 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.365 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:42.365 "name": "Existed_Raid", 00:15:42.365 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.365 "strip_size_kb": 64, 00:15:42.365 "state": "configuring", 00:15:42.365 "raid_level": "raid5f", 00:15:42.365 "superblock": false, 00:15:42.365 "num_base_bdevs": 4, 00:15:42.365 "num_base_bdevs_discovered": 1, 00:15:42.365 "num_base_bdevs_operational": 4, 00:15:42.365 "base_bdevs_list": [ 00:15:42.365 { 00:15:42.365 "name": "BaseBdev1", 00:15:42.365 "uuid": "23a1a968-4da8-4d6c-87a5-c96cdcbf0898", 00:15:42.365 "is_configured": true, 00:15:42.365 "data_offset": 0, 00:15:42.365 "data_size": 65536 00:15:42.365 }, 00:15:42.365 { 00:15:42.365 "name": "BaseBdev2", 00:15:42.365 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.365 "is_configured": false, 00:15:42.365 "data_offset": 0, 00:15:42.365 "data_size": 0 00:15:42.365 }, 00:15:42.365 { 00:15:42.365 "name": "BaseBdev3", 00:15:42.365 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.365 "is_configured": false, 00:15:42.365 "data_offset": 0, 00:15:42.365 "data_size": 0 00:15:42.365 }, 00:15:42.365 { 00:15:42.365 "name": "BaseBdev4", 00:15:42.365 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.365 "is_configured": false, 00:15:42.365 "data_offset": 0, 00:15:42.365 "data_size": 0 00:15:42.365 } 00:15:42.365 ] 00:15:42.365 }' 00:15:42.365 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:42.365 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.626 [2024-11-20 16:03:40.767003] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:42.626 BaseBdev2 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.626 [ 00:15:42.626 { 00:15:42.626 "name": "BaseBdev2", 00:15:42.626 "aliases": [ 00:15:42.626 "6f3c9620-e115-4fe3-a68b-5073d1765b3c" 00:15:42.626 ], 00:15:42.626 "product_name": "Malloc disk", 00:15:42.626 "block_size": 512, 00:15:42.626 "num_blocks": 65536, 00:15:42.626 "uuid": "6f3c9620-e115-4fe3-a68b-5073d1765b3c", 00:15:42.626 "assigned_rate_limits": { 00:15:42.626 "rw_ios_per_sec": 0, 00:15:42.626 "rw_mbytes_per_sec": 0, 00:15:42.626 "r_mbytes_per_sec": 0, 00:15:42.626 "w_mbytes_per_sec": 0 00:15:42.626 }, 00:15:42.626 "claimed": true, 00:15:42.626 "claim_type": "exclusive_write", 00:15:42.626 "zoned": false, 00:15:42.626 "supported_io_types": { 00:15:42.626 "read": true, 00:15:42.626 "write": true, 00:15:42.626 "unmap": true, 00:15:42.626 "flush": true, 00:15:42.626 "reset": true, 00:15:42.626 "nvme_admin": false, 00:15:42.626 "nvme_io": false, 00:15:42.626 "nvme_io_md": false, 00:15:42.626 "write_zeroes": true, 00:15:42.626 "zcopy": true, 00:15:42.626 "get_zone_info": false, 00:15:42.626 "zone_management": false, 00:15:42.626 "zone_append": false, 00:15:42.626 "compare": false, 00:15:42.626 "compare_and_write": false, 00:15:42.626 "abort": true, 00:15:42.626 "seek_hole": false, 00:15:42.626 "seek_data": false, 00:15:42.626 "copy": true, 00:15:42.626 "nvme_iov_md": false 00:15:42.626 }, 00:15:42.626 "memory_domains": [ 00:15:42.626 { 00:15:42.626 "dma_device_id": "system", 00:15:42.626 "dma_device_type": 1 00:15:42.626 }, 00:15:42.626 { 00:15:42.626 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:42.626 "dma_device_type": 2 00:15:42.626 } 00:15:42.626 ], 00:15:42.626 "driver_specific": {} 00:15:42.626 } 00:15:42.626 ] 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:42.626 "name": "Existed_Raid", 00:15:42.626 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.626 "strip_size_kb": 64, 00:15:42.626 "state": "configuring", 00:15:42.626 "raid_level": "raid5f", 00:15:42.626 "superblock": false, 00:15:42.626 "num_base_bdevs": 4, 00:15:42.626 "num_base_bdevs_discovered": 2, 00:15:42.626 "num_base_bdevs_operational": 4, 00:15:42.626 "base_bdevs_list": [ 00:15:42.626 { 00:15:42.626 "name": "BaseBdev1", 00:15:42.626 "uuid": "23a1a968-4da8-4d6c-87a5-c96cdcbf0898", 00:15:42.626 "is_configured": true, 00:15:42.626 "data_offset": 0, 00:15:42.626 "data_size": 65536 00:15:42.626 }, 00:15:42.626 { 00:15:42.626 "name": "BaseBdev2", 00:15:42.626 "uuid": "6f3c9620-e115-4fe3-a68b-5073d1765b3c", 00:15:42.626 "is_configured": true, 00:15:42.626 "data_offset": 0, 00:15:42.626 "data_size": 65536 00:15:42.626 }, 00:15:42.626 { 00:15:42.626 "name": "BaseBdev3", 00:15:42.626 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.626 "is_configured": false, 00:15:42.626 "data_offset": 0, 00:15:42.626 "data_size": 0 00:15:42.626 }, 00:15:42.626 { 00:15:42.626 "name": "BaseBdev4", 00:15:42.626 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.626 "is_configured": false, 00:15:42.626 "data_offset": 0, 00:15:42.626 "data_size": 0 00:15:42.626 } 00:15:42.626 ] 00:15:42.626 }' 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:42.626 16:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.886 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:15:42.886 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.886 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.148 [2024-11-20 16:03:41.161111] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:43.148 BaseBdev3 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.148 [ 00:15:43.148 { 00:15:43.148 "name": "BaseBdev3", 00:15:43.148 "aliases": [ 00:15:43.148 "b364f2c3-0de9-4aa5-b48d-4e762e699775" 00:15:43.148 ], 00:15:43.148 "product_name": "Malloc disk", 00:15:43.148 "block_size": 512, 00:15:43.148 "num_blocks": 65536, 00:15:43.148 "uuid": "b364f2c3-0de9-4aa5-b48d-4e762e699775", 00:15:43.148 "assigned_rate_limits": { 00:15:43.148 "rw_ios_per_sec": 0, 00:15:43.148 "rw_mbytes_per_sec": 0, 00:15:43.148 "r_mbytes_per_sec": 0, 00:15:43.148 "w_mbytes_per_sec": 0 00:15:43.148 }, 00:15:43.148 "claimed": true, 00:15:43.148 "claim_type": "exclusive_write", 00:15:43.148 "zoned": false, 00:15:43.148 "supported_io_types": { 00:15:43.148 "read": true, 00:15:43.148 "write": true, 00:15:43.148 "unmap": true, 00:15:43.148 "flush": true, 00:15:43.148 "reset": true, 00:15:43.148 "nvme_admin": false, 00:15:43.148 "nvme_io": false, 00:15:43.148 "nvme_io_md": false, 00:15:43.148 "write_zeroes": true, 00:15:43.148 "zcopy": true, 00:15:43.148 "get_zone_info": false, 00:15:43.148 "zone_management": false, 00:15:43.148 "zone_append": false, 00:15:43.148 "compare": false, 00:15:43.148 "compare_and_write": false, 00:15:43.148 "abort": true, 00:15:43.148 "seek_hole": false, 00:15:43.148 "seek_data": false, 00:15:43.148 "copy": true, 00:15:43.148 "nvme_iov_md": false 00:15:43.148 }, 00:15:43.148 "memory_domains": [ 00:15:43.148 { 00:15:43.148 "dma_device_id": "system", 00:15:43.148 "dma_device_type": 1 00:15:43.148 }, 00:15:43.148 { 00:15:43.148 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:43.148 "dma_device_type": 2 00:15:43.148 } 00:15:43.148 ], 00:15:43.148 "driver_specific": {} 00:15:43.148 } 00:15:43.148 ] 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:43.148 "name": "Existed_Raid", 00:15:43.148 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:43.148 "strip_size_kb": 64, 00:15:43.148 "state": "configuring", 00:15:43.148 "raid_level": "raid5f", 00:15:43.148 "superblock": false, 00:15:43.148 "num_base_bdevs": 4, 00:15:43.148 "num_base_bdevs_discovered": 3, 00:15:43.148 "num_base_bdevs_operational": 4, 00:15:43.148 "base_bdevs_list": [ 00:15:43.148 { 00:15:43.148 "name": "BaseBdev1", 00:15:43.148 "uuid": "23a1a968-4da8-4d6c-87a5-c96cdcbf0898", 00:15:43.148 "is_configured": true, 00:15:43.148 "data_offset": 0, 00:15:43.148 "data_size": 65536 00:15:43.148 }, 00:15:43.148 { 00:15:43.148 "name": "BaseBdev2", 00:15:43.148 "uuid": "6f3c9620-e115-4fe3-a68b-5073d1765b3c", 00:15:43.148 "is_configured": true, 00:15:43.148 "data_offset": 0, 00:15:43.148 "data_size": 65536 00:15:43.148 }, 00:15:43.148 { 00:15:43.148 "name": "BaseBdev3", 00:15:43.148 "uuid": "b364f2c3-0de9-4aa5-b48d-4e762e699775", 00:15:43.148 "is_configured": true, 00:15:43.148 "data_offset": 0, 00:15:43.148 "data_size": 65536 00:15:43.148 }, 00:15:43.148 { 00:15:43.148 "name": "BaseBdev4", 00:15:43.148 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:43.148 "is_configured": false, 00:15:43.148 "data_offset": 0, 00:15:43.148 "data_size": 0 00:15:43.148 } 00:15:43.148 ] 00:15:43.148 }' 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:43.148 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.409 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:15:43.409 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.409 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.409 [2024-11-20 16:03:41.532051] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:43.409 [2024-11-20 16:03:41.532103] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:15:43.409 [2024-11-20 16:03:41.532115] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:15:43.409 [2024-11-20 16:03:41.532369] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:15:43.409 [2024-11-20 16:03:41.537305] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:15:43.409 [2024-11-20 16:03:41.537326] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:15:43.409 [2024-11-20 16:03:41.537565] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:43.409 BaseBdev4 00:15:43.409 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.409 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:15:43.409 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:15:43.409 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:43.409 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:43.409 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:43.409 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:43.409 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:43.409 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.409 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.409 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.409 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:15:43.409 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.409 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.409 [ 00:15:43.409 { 00:15:43.409 "name": "BaseBdev4", 00:15:43.409 "aliases": [ 00:15:43.409 "16f532fa-4f59-4e90-a9f9-4e3fa7fe2b4e" 00:15:43.409 ], 00:15:43.409 "product_name": "Malloc disk", 00:15:43.409 "block_size": 512, 00:15:43.409 "num_blocks": 65536, 00:15:43.409 "uuid": "16f532fa-4f59-4e90-a9f9-4e3fa7fe2b4e", 00:15:43.409 "assigned_rate_limits": { 00:15:43.409 "rw_ios_per_sec": 0, 00:15:43.409 "rw_mbytes_per_sec": 0, 00:15:43.409 "r_mbytes_per_sec": 0, 00:15:43.409 "w_mbytes_per_sec": 0 00:15:43.409 }, 00:15:43.409 "claimed": true, 00:15:43.409 "claim_type": "exclusive_write", 00:15:43.409 "zoned": false, 00:15:43.409 "supported_io_types": { 00:15:43.409 "read": true, 00:15:43.409 "write": true, 00:15:43.409 "unmap": true, 00:15:43.409 "flush": true, 00:15:43.409 "reset": true, 00:15:43.410 "nvme_admin": false, 00:15:43.410 "nvme_io": false, 00:15:43.410 "nvme_io_md": false, 00:15:43.410 "write_zeroes": true, 00:15:43.410 "zcopy": true, 00:15:43.410 "get_zone_info": false, 00:15:43.410 "zone_management": false, 00:15:43.410 "zone_append": false, 00:15:43.410 "compare": false, 00:15:43.410 "compare_and_write": false, 00:15:43.410 "abort": true, 00:15:43.410 "seek_hole": false, 00:15:43.410 "seek_data": false, 00:15:43.410 "copy": true, 00:15:43.410 "nvme_iov_md": false 00:15:43.410 }, 00:15:43.410 "memory_domains": [ 00:15:43.410 { 00:15:43.410 "dma_device_id": "system", 00:15:43.410 "dma_device_type": 1 00:15:43.410 }, 00:15:43.410 { 00:15:43.410 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:43.410 "dma_device_type": 2 00:15:43.410 } 00:15:43.410 ], 00:15:43.410 "driver_specific": {} 00:15:43.410 } 00:15:43.410 ] 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:43.410 "name": "Existed_Raid", 00:15:43.410 "uuid": "ab5e21bf-56b8-499e-9548-35b70cf47c3d", 00:15:43.410 "strip_size_kb": 64, 00:15:43.410 "state": "online", 00:15:43.410 "raid_level": "raid5f", 00:15:43.410 "superblock": false, 00:15:43.410 "num_base_bdevs": 4, 00:15:43.410 "num_base_bdevs_discovered": 4, 00:15:43.410 "num_base_bdevs_operational": 4, 00:15:43.410 "base_bdevs_list": [ 00:15:43.410 { 00:15:43.410 "name": "BaseBdev1", 00:15:43.410 "uuid": "23a1a968-4da8-4d6c-87a5-c96cdcbf0898", 00:15:43.410 "is_configured": true, 00:15:43.410 "data_offset": 0, 00:15:43.410 "data_size": 65536 00:15:43.410 }, 00:15:43.410 { 00:15:43.410 "name": "BaseBdev2", 00:15:43.410 "uuid": "6f3c9620-e115-4fe3-a68b-5073d1765b3c", 00:15:43.410 "is_configured": true, 00:15:43.410 "data_offset": 0, 00:15:43.410 "data_size": 65536 00:15:43.410 }, 00:15:43.410 { 00:15:43.410 "name": "BaseBdev3", 00:15:43.410 "uuid": "b364f2c3-0de9-4aa5-b48d-4e762e699775", 00:15:43.410 "is_configured": true, 00:15:43.410 "data_offset": 0, 00:15:43.410 "data_size": 65536 00:15:43.410 }, 00:15:43.410 { 00:15:43.410 "name": "BaseBdev4", 00:15:43.410 "uuid": "16f532fa-4f59-4e90-a9f9-4e3fa7fe2b4e", 00:15:43.410 "is_configured": true, 00:15:43.410 "data_offset": 0, 00:15:43.410 "data_size": 65536 00:15:43.410 } 00:15:43.410 ] 00:15:43.410 }' 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:43.410 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.670 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:15:43.670 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:43.670 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:43.670 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:43.670 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:15:43.670 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:43.670 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:43.670 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.670 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.670 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:43.670 [2024-11-20 16:03:41.883688] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:43.670 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.670 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:43.670 "name": "Existed_Raid", 00:15:43.670 "aliases": [ 00:15:43.670 "ab5e21bf-56b8-499e-9548-35b70cf47c3d" 00:15:43.670 ], 00:15:43.670 "product_name": "Raid Volume", 00:15:43.670 "block_size": 512, 00:15:43.670 "num_blocks": 196608, 00:15:43.670 "uuid": "ab5e21bf-56b8-499e-9548-35b70cf47c3d", 00:15:43.670 "assigned_rate_limits": { 00:15:43.670 "rw_ios_per_sec": 0, 00:15:43.670 "rw_mbytes_per_sec": 0, 00:15:43.670 "r_mbytes_per_sec": 0, 00:15:43.670 "w_mbytes_per_sec": 0 00:15:43.670 }, 00:15:43.670 "claimed": false, 00:15:43.670 "zoned": false, 00:15:43.670 "supported_io_types": { 00:15:43.670 "read": true, 00:15:43.670 "write": true, 00:15:43.670 "unmap": false, 00:15:43.670 "flush": false, 00:15:43.670 "reset": true, 00:15:43.670 "nvme_admin": false, 00:15:43.670 "nvme_io": false, 00:15:43.670 "nvme_io_md": false, 00:15:43.670 "write_zeroes": true, 00:15:43.670 "zcopy": false, 00:15:43.670 "get_zone_info": false, 00:15:43.670 "zone_management": false, 00:15:43.670 "zone_append": false, 00:15:43.670 "compare": false, 00:15:43.670 "compare_and_write": false, 00:15:43.670 "abort": false, 00:15:43.670 "seek_hole": false, 00:15:43.670 "seek_data": false, 00:15:43.670 "copy": false, 00:15:43.670 "nvme_iov_md": false 00:15:43.670 }, 00:15:43.670 "driver_specific": { 00:15:43.670 "raid": { 00:15:43.670 "uuid": "ab5e21bf-56b8-499e-9548-35b70cf47c3d", 00:15:43.670 "strip_size_kb": 64, 00:15:43.670 "state": "online", 00:15:43.670 "raid_level": "raid5f", 00:15:43.671 "superblock": false, 00:15:43.671 "num_base_bdevs": 4, 00:15:43.671 "num_base_bdevs_discovered": 4, 00:15:43.671 "num_base_bdevs_operational": 4, 00:15:43.671 "base_bdevs_list": [ 00:15:43.671 { 00:15:43.671 "name": "BaseBdev1", 00:15:43.671 "uuid": "23a1a968-4da8-4d6c-87a5-c96cdcbf0898", 00:15:43.671 "is_configured": true, 00:15:43.671 "data_offset": 0, 00:15:43.671 "data_size": 65536 00:15:43.671 }, 00:15:43.671 { 00:15:43.671 "name": "BaseBdev2", 00:15:43.671 "uuid": "6f3c9620-e115-4fe3-a68b-5073d1765b3c", 00:15:43.671 "is_configured": true, 00:15:43.671 "data_offset": 0, 00:15:43.671 "data_size": 65536 00:15:43.671 }, 00:15:43.671 { 00:15:43.671 "name": "BaseBdev3", 00:15:43.671 "uuid": "b364f2c3-0de9-4aa5-b48d-4e762e699775", 00:15:43.671 "is_configured": true, 00:15:43.671 "data_offset": 0, 00:15:43.671 "data_size": 65536 00:15:43.671 }, 00:15:43.671 { 00:15:43.671 "name": "BaseBdev4", 00:15:43.671 "uuid": "16f532fa-4f59-4e90-a9f9-4e3fa7fe2b4e", 00:15:43.671 "is_configured": true, 00:15:43.671 "data_offset": 0, 00:15:43.671 "data_size": 65536 00:15:43.671 } 00:15:43.671 ] 00:15:43.671 } 00:15:43.671 } 00:15:43.671 }' 00:15:43.671 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:43.930 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:15:43.930 BaseBdev2 00:15:43.930 BaseBdev3 00:15:43.930 BaseBdev4' 00:15:43.930 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:43.930 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:43.930 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:43.930 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:15:43.930 16:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:43.930 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.930 16:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.930 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.930 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:43.930 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:43.930 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:43.930 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:43.930 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.931 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.931 [2024-11-20 16:03:42.131574] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:44.192 "name": "Existed_Raid", 00:15:44.192 "uuid": "ab5e21bf-56b8-499e-9548-35b70cf47c3d", 00:15:44.192 "strip_size_kb": 64, 00:15:44.192 "state": "online", 00:15:44.192 "raid_level": "raid5f", 00:15:44.192 "superblock": false, 00:15:44.192 "num_base_bdevs": 4, 00:15:44.192 "num_base_bdevs_discovered": 3, 00:15:44.192 "num_base_bdevs_operational": 3, 00:15:44.192 "base_bdevs_list": [ 00:15:44.192 { 00:15:44.192 "name": null, 00:15:44.192 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:44.192 "is_configured": false, 00:15:44.192 "data_offset": 0, 00:15:44.192 "data_size": 65536 00:15:44.192 }, 00:15:44.192 { 00:15:44.192 "name": "BaseBdev2", 00:15:44.192 "uuid": "6f3c9620-e115-4fe3-a68b-5073d1765b3c", 00:15:44.192 "is_configured": true, 00:15:44.192 "data_offset": 0, 00:15:44.192 "data_size": 65536 00:15:44.192 }, 00:15:44.192 { 00:15:44.192 "name": "BaseBdev3", 00:15:44.192 "uuid": "b364f2c3-0de9-4aa5-b48d-4e762e699775", 00:15:44.192 "is_configured": true, 00:15:44.192 "data_offset": 0, 00:15:44.192 "data_size": 65536 00:15:44.192 }, 00:15:44.192 { 00:15:44.192 "name": "BaseBdev4", 00:15:44.192 "uuid": "16f532fa-4f59-4e90-a9f9-4e3fa7fe2b4e", 00:15:44.192 "is_configured": true, 00:15:44.192 "data_offset": 0, 00:15:44.192 "data_size": 65536 00:15:44.192 } 00:15:44.192 ] 00:15:44.192 }' 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:44.192 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.453 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:15:44.453 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:44.453 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.453 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:44.453 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.453 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.453 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.453 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:44.453 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:44.453 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:15:44.453 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.453 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.453 [2024-11-20 16:03:42.530543] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:44.453 [2024-11-20 16:03:42.530739] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:44.453 [2024-11-20 16:03:42.588440] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:44.453 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.453 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:44.453 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:44.453 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.454 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:44.454 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.454 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.454 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.454 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:44.454 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:44.454 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:15:44.454 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.454 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.454 [2024-11-20 16:03:42.628484] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:15:44.454 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.454 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:44.454 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:44.454 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:44.454 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.454 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.454 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.714 [2024-11-20 16:03:42.730878] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:15:44.714 [2024-11-20 16:03:42.731023] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.714 BaseBdev2 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.714 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.714 [ 00:15:44.714 { 00:15:44.714 "name": "BaseBdev2", 00:15:44.714 "aliases": [ 00:15:44.714 "26e7d4b2-a0d4-4d7e-957e-26073e3387bc" 00:15:44.714 ], 00:15:44.714 "product_name": "Malloc disk", 00:15:44.714 "block_size": 512, 00:15:44.714 "num_blocks": 65536, 00:15:44.714 "uuid": "26e7d4b2-a0d4-4d7e-957e-26073e3387bc", 00:15:44.714 "assigned_rate_limits": { 00:15:44.714 "rw_ios_per_sec": 0, 00:15:44.714 "rw_mbytes_per_sec": 0, 00:15:44.714 "r_mbytes_per_sec": 0, 00:15:44.714 "w_mbytes_per_sec": 0 00:15:44.714 }, 00:15:44.714 "claimed": false, 00:15:44.714 "zoned": false, 00:15:44.714 "supported_io_types": { 00:15:44.714 "read": true, 00:15:44.714 "write": true, 00:15:44.714 "unmap": true, 00:15:44.714 "flush": true, 00:15:44.714 "reset": true, 00:15:44.714 "nvme_admin": false, 00:15:44.714 "nvme_io": false, 00:15:44.714 "nvme_io_md": false, 00:15:44.714 "write_zeroes": true, 00:15:44.714 "zcopy": true, 00:15:44.714 "get_zone_info": false, 00:15:44.714 "zone_management": false, 00:15:44.714 "zone_append": false, 00:15:44.715 "compare": false, 00:15:44.715 "compare_and_write": false, 00:15:44.715 "abort": true, 00:15:44.715 "seek_hole": false, 00:15:44.715 "seek_data": false, 00:15:44.715 "copy": true, 00:15:44.715 "nvme_iov_md": false 00:15:44.715 }, 00:15:44.715 "memory_domains": [ 00:15:44.715 { 00:15:44.715 "dma_device_id": "system", 00:15:44.715 "dma_device_type": 1 00:15:44.715 }, 00:15:44.715 { 00:15:44.715 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:44.715 "dma_device_type": 2 00:15:44.715 } 00:15:44.715 ], 00:15:44.715 "driver_specific": {} 00:15:44.715 } 00:15:44.715 ] 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.715 BaseBdev3 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.715 [ 00:15:44.715 { 00:15:44.715 "name": "BaseBdev3", 00:15:44.715 "aliases": [ 00:15:44.715 "c7789c72-78b4-42dd-8ea7-ecaa7a28a5d8" 00:15:44.715 ], 00:15:44.715 "product_name": "Malloc disk", 00:15:44.715 "block_size": 512, 00:15:44.715 "num_blocks": 65536, 00:15:44.715 "uuid": "c7789c72-78b4-42dd-8ea7-ecaa7a28a5d8", 00:15:44.715 "assigned_rate_limits": { 00:15:44.715 "rw_ios_per_sec": 0, 00:15:44.715 "rw_mbytes_per_sec": 0, 00:15:44.715 "r_mbytes_per_sec": 0, 00:15:44.715 "w_mbytes_per_sec": 0 00:15:44.715 }, 00:15:44.715 "claimed": false, 00:15:44.715 "zoned": false, 00:15:44.715 "supported_io_types": { 00:15:44.715 "read": true, 00:15:44.715 "write": true, 00:15:44.715 "unmap": true, 00:15:44.715 "flush": true, 00:15:44.715 "reset": true, 00:15:44.715 "nvme_admin": false, 00:15:44.715 "nvme_io": false, 00:15:44.715 "nvme_io_md": false, 00:15:44.715 "write_zeroes": true, 00:15:44.715 "zcopy": true, 00:15:44.715 "get_zone_info": false, 00:15:44.715 "zone_management": false, 00:15:44.715 "zone_append": false, 00:15:44.715 "compare": false, 00:15:44.715 "compare_and_write": false, 00:15:44.715 "abort": true, 00:15:44.715 "seek_hole": false, 00:15:44.715 "seek_data": false, 00:15:44.715 "copy": true, 00:15:44.715 "nvme_iov_md": false 00:15:44.715 }, 00:15:44.715 "memory_domains": [ 00:15:44.715 { 00:15:44.715 "dma_device_id": "system", 00:15:44.715 "dma_device_type": 1 00:15:44.715 }, 00:15:44.715 { 00:15:44.715 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:44.715 "dma_device_type": 2 00:15:44.715 } 00:15:44.715 ], 00:15:44.715 "driver_specific": {} 00:15:44.715 } 00:15:44.715 ] 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.715 BaseBdev4 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.715 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.975 [ 00:15:44.975 { 00:15:44.975 "name": "BaseBdev4", 00:15:44.975 "aliases": [ 00:15:44.975 "2f968fda-716a-4d74-a09c-141e9728493b" 00:15:44.975 ], 00:15:44.975 "product_name": "Malloc disk", 00:15:44.975 "block_size": 512, 00:15:44.975 "num_blocks": 65536, 00:15:44.975 "uuid": "2f968fda-716a-4d74-a09c-141e9728493b", 00:15:44.975 "assigned_rate_limits": { 00:15:44.975 "rw_ios_per_sec": 0, 00:15:44.975 "rw_mbytes_per_sec": 0, 00:15:44.975 "r_mbytes_per_sec": 0, 00:15:44.975 "w_mbytes_per_sec": 0 00:15:44.975 }, 00:15:44.975 "claimed": false, 00:15:44.975 "zoned": false, 00:15:44.975 "supported_io_types": { 00:15:44.975 "read": true, 00:15:44.975 "write": true, 00:15:44.975 "unmap": true, 00:15:44.975 "flush": true, 00:15:44.975 "reset": true, 00:15:44.975 "nvme_admin": false, 00:15:44.975 "nvme_io": false, 00:15:44.975 "nvme_io_md": false, 00:15:44.975 "write_zeroes": true, 00:15:44.975 "zcopy": true, 00:15:44.975 "get_zone_info": false, 00:15:44.975 "zone_management": false, 00:15:44.975 "zone_append": false, 00:15:44.975 "compare": false, 00:15:44.975 "compare_and_write": false, 00:15:44.975 "abort": true, 00:15:44.975 "seek_hole": false, 00:15:44.975 "seek_data": false, 00:15:44.975 "copy": true, 00:15:44.975 "nvme_iov_md": false 00:15:44.975 }, 00:15:44.975 "memory_domains": [ 00:15:44.975 { 00:15:44.975 "dma_device_id": "system", 00:15:44.975 "dma_device_type": 1 00:15:44.975 }, 00:15:44.975 { 00:15:44.975 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:44.975 "dma_device_type": 2 00:15:44.975 } 00:15:44.975 ], 00:15:44.975 "driver_specific": {} 00:15:44.975 } 00:15:44.975 ] 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.975 [2024-11-20 16:03:42.986237] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:44.975 [2024-11-20 16:03:42.986370] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:44.975 [2024-11-20 16:03:42.986440] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:44.975 [2024-11-20 16:03:42.988305] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:44.975 [2024-11-20 16:03:42.988433] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.975 16:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.975 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.975 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:44.975 "name": "Existed_Raid", 00:15:44.975 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:44.975 "strip_size_kb": 64, 00:15:44.975 "state": "configuring", 00:15:44.975 "raid_level": "raid5f", 00:15:44.975 "superblock": false, 00:15:44.975 "num_base_bdevs": 4, 00:15:44.975 "num_base_bdevs_discovered": 3, 00:15:44.975 "num_base_bdevs_operational": 4, 00:15:44.975 "base_bdevs_list": [ 00:15:44.975 { 00:15:44.975 "name": "BaseBdev1", 00:15:44.975 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:44.975 "is_configured": false, 00:15:44.975 "data_offset": 0, 00:15:44.975 "data_size": 0 00:15:44.975 }, 00:15:44.975 { 00:15:44.975 "name": "BaseBdev2", 00:15:44.975 "uuid": "26e7d4b2-a0d4-4d7e-957e-26073e3387bc", 00:15:44.975 "is_configured": true, 00:15:44.975 "data_offset": 0, 00:15:44.975 "data_size": 65536 00:15:44.975 }, 00:15:44.975 { 00:15:44.975 "name": "BaseBdev3", 00:15:44.975 "uuid": "c7789c72-78b4-42dd-8ea7-ecaa7a28a5d8", 00:15:44.975 "is_configured": true, 00:15:44.975 "data_offset": 0, 00:15:44.975 "data_size": 65536 00:15:44.975 }, 00:15:44.975 { 00:15:44.975 "name": "BaseBdev4", 00:15:44.975 "uuid": "2f968fda-716a-4d74-a09c-141e9728493b", 00:15:44.975 "is_configured": true, 00:15:44.975 "data_offset": 0, 00:15:44.975 "data_size": 65536 00:15:44.975 } 00:15:44.975 ] 00:15:44.975 }' 00:15:44.975 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:44.975 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.235 [2024-11-20 16:03:43.294322] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:45.235 "name": "Existed_Raid", 00:15:45.235 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:45.235 "strip_size_kb": 64, 00:15:45.235 "state": "configuring", 00:15:45.235 "raid_level": "raid5f", 00:15:45.235 "superblock": false, 00:15:45.235 "num_base_bdevs": 4, 00:15:45.235 "num_base_bdevs_discovered": 2, 00:15:45.235 "num_base_bdevs_operational": 4, 00:15:45.235 "base_bdevs_list": [ 00:15:45.235 { 00:15:45.235 "name": "BaseBdev1", 00:15:45.235 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:45.235 "is_configured": false, 00:15:45.235 "data_offset": 0, 00:15:45.235 "data_size": 0 00:15:45.235 }, 00:15:45.235 { 00:15:45.235 "name": null, 00:15:45.235 "uuid": "26e7d4b2-a0d4-4d7e-957e-26073e3387bc", 00:15:45.235 "is_configured": false, 00:15:45.235 "data_offset": 0, 00:15:45.235 "data_size": 65536 00:15:45.235 }, 00:15:45.235 { 00:15:45.235 "name": "BaseBdev3", 00:15:45.235 "uuid": "c7789c72-78b4-42dd-8ea7-ecaa7a28a5d8", 00:15:45.235 "is_configured": true, 00:15:45.235 "data_offset": 0, 00:15:45.235 "data_size": 65536 00:15:45.235 }, 00:15:45.235 { 00:15:45.235 "name": "BaseBdev4", 00:15:45.235 "uuid": "2f968fda-716a-4d74-a09c-141e9728493b", 00:15:45.235 "is_configured": true, 00:15:45.235 "data_offset": 0, 00:15:45.235 "data_size": 65536 00:15:45.235 } 00:15:45.235 ] 00:15:45.235 }' 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:45.235 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.497 [2024-11-20 16:03:43.709072] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:45.497 BaseBdev1 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.497 [ 00:15:45.497 { 00:15:45.497 "name": "BaseBdev1", 00:15:45.497 "aliases": [ 00:15:45.497 "9cb206a8-bb09-4f66-9b2c-04a931ac5437" 00:15:45.497 ], 00:15:45.497 "product_name": "Malloc disk", 00:15:45.497 "block_size": 512, 00:15:45.497 "num_blocks": 65536, 00:15:45.497 "uuid": "9cb206a8-bb09-4f66-9b2c-04a931ac5437", 00:15:45.497 "assigned_rate_limits": { 00:15:45.497 "rw_ios_per_sec": 0, 00:15:45.497 "rw_mbytes_per_sec": 0, 00:15:45.497 "r_mbytes_per_sec": 0, 00:15:45.497 "w_mbytes_per_sec": 0 00:15:45.497 }, 00:15:45.497 "claimed": true, 00:15:45.497 "claim_type": "exclusive_write", 00:15:45.497 "zoned": false, 00:15:45.497 "supported_io_types": { 00:15:45.497 "read": true, 00:15:45.497 "write": true, 00:15:45.497 "unmap": true, 00:15:45.497 "flush": true, 00:15:45.497 "reset": true, 00:15:45.497 "nvme_admin": false, 00:15:45.497 "nvme_io": false, 00:15:45.497 "nvme_io_md": false, 00:15:45.497 "write_zeroes": true, 00:15:45.497 "zcopy": true, 00:15:45.497 "get_zone_info": false, 00:15:45.497 "zone_management": false, 00:15:45.497 "zone_append": false, 00:15:45.497 "compare": false, 00:15:45.497 "compare_and_write": false, 00:15:45.497 "abort": true, 00:15:45.497 "seek_hole": false, 00:15:45.497 "seek_data": false, 00:15:45.497 "copy": true, 00:15:45.497 "nvme_iov_md": false 00:15:45.497 }, 00:15:45.497 "memory_domains": [ 00:15:45.497 { 00:15:45.497 "dma_device_id": "system", 00:15:45.497 "dma_device_type": 1 00:15:45.497 }, 00:15:45.497 { 00:15:45.497 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:45.497 "dma_device_type": 2 00:15:45.497 } 00:15:45.497 ], 00:15:45.497 "driver_specific": {} 00:15:45.497 } 00:15:45.497 ] 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.497 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.759 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.759 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:45.759 "name": "Existed_Raid", 00:15:45.759 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:45.759 "strip_size_kb": 64, 00:15:45.759 "state": "configuring", 00:15:45.759 "raid_level": "raid5f", 00:15:45.759 "superblock": false, 00:15:45.759 "num_base_bdevs": 4, 00:15:45.759 "num_base_bdevs_discovered": 3, 00:15:45.759 "num_base_bdevs_operational": 4, 00:15:45.759 "base_bdevs_list": [ 00:15:45.759 { 00:15:45.759 "name": "BaseBdev1", 00:15:45.759 "uuid": "9cb206a8-bb09-4f66-9b2c-04a931ac5437", 00:15:45.759 "is_configured": true, 00:15:45.759 "data_offset": 0, 00:15:45.759 "data_size": 65536 00:15:45.759 }, 00:15:45.759 { 00:15:45.759 "name": null, 00:15:45.759 "uuid": "26e7d4b2-a0d4-4d7e-957e-26073e3387bc", 00:15:45.759 "is_configured": false, 00:15:45.759 "data_offset": 0, 00:15:45.759 "data_size": 65536 00:15:45.759 }, 00:15:45.759 { 00:15:45.759 "name": "BaseBdev3", 00:15:45.759 "uuid": "c7789c72-78b4-42dd-8ea7-ecaa7a28a5d8", 00:15:45.759 "is_configured": true, 00:15:45.759 "data_offset": 0, 00:15:45.759 "data_size": 65536 00:15:45.759 }, 00:15:45.759 { 00:15:45.759 "name": "BaseBdev4", 00:15:45.759 "uuid": "2f968fda-716a-4d74-a09c-141e9728493b", 00:15:45.759 "is_configured": true, 00:15:45.759 "data_offset": 0, 00:15:45.759 "data_size": 65536 00:15:45.759 } 00:15:45.759 ] 00:15:45.759 }' 00:15:45.759 16:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:45.759 16:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.020 [2024-11-20 16:03:44.101254] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:46.020 "name": "Existed_Raid", 00:15:46.020 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:46.020 "strip_size_kb": 64, 00:15:46.020 "state": "configuring", 00:15:46.020 "raid_level": "raid5f", 00:15:46.020 "superblock": false, 00:15:46.020 "num_base_bdevs": 4, 00:15:46.020 "num_base_bdevs_discovered": 2, 00:15:46.020 "num_base_bdevs_operational": 4, 00:15:46.020 "base_bdevs_list": [ 00:15:46.020 { 00:15:46.020 "name": "BaseBdev1", 00:15:46.020 "uuid": "9cb206a8-bb09-4f66-9b2c-04a931ac5437", 00:15:46.020 "is_configured": true, 00:15:46.020 "data_offset": 0, 00:15:46.020 "data_size": 65536 00:15:46.020 }, 00:15:46.020 { 00:15:46.020 "name": null, 00:15:46.020 "uuid": "26e7d4b2-a0d4-4d7e-957e-26073e3387bc", 00:15:46.020 "is_configured": false, 00:15:46.020 "data_offset": 0, 00:15:46.020 "data_size": 65536 00:15:46.020 }, 00:15:46.020 { 00:15:46.020 "name": null, 00:15:46.020 "uuid": "c7789c72-78b4-42dd-8ea7-ecaa7a28a5d8", 00:15:46.020 "is_configured": false, 00:15:46.020 "data_offset": 0, 00:15:46.020 "data_size": 65536 00:15:46.020 }, 00:15:46.020 { 00:15:46.020 "name": "BaseBdev4", 00:15:46.020 "uuid": "2f968fda-716a-4d74-a09c-141e9728493b", 00:15:46.020 "is_configured": true, 00:15:46.020 "data_offset": 0, 00:15:46.020 "data_size": 65536 00:15:46.020 } 00:15:46.020 ] 00:15:46.020 }' 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:46.020 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.280 [2024-11-20 16:03:44.441321] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.280 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:46.280 "name": "Existed_Raid", 00:15:46.280 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:46.280 "strip_size_kb": 64, 00:15:46.280 "state": "configuring", 00:15:46.280 "raid_level": "raid5f", 00:15:46.280 "superblock": false, 00:15:46.280 "num_base_bdevs": 4, 00:15:46.280 "num_base_bdevs_discovered": 3, 00:15:46.280 "num_base_bdevs_operational": 4, 00:15:46.280 "base_bdevs_list": [ 00:15:46.280 { 00:15:46.280 "name": "BaseBdev1", 00:15:46.280 "uuid": "9cb206a8-bb09-4f66-9b2c-04a931ac5437", 00:15:46.280 "is_configured": true, 00:15:46.280 "data_offset": 0, 00:15:46.280 "data_size": 65536 00:15:46.281 }, 00:15:46.281 { 00:15:46.281 "name": null, 00:15:46.281 "uuid": "26e7d4b2-a0d4-4d7e-957e-26073e3387bc", 00:15:46.281 "is_configured": false, 00:15:46.281 "data_offset": 0, 00:15:46.281 "data_size": 65536 00:15:46.281 }, 00:15:46.281 { 00:15:46.281 "name": "BaseBdev3", 00:15:46.281 "uuid": "c7789c72-78b4-42dd-8ea7-ecaa7a28a5d8", 00:15:46.281 "is_configured": true, 00:15:46.281 "data_offset": 0, 00:15:46.281 "data_size": 65536 00:15:46.281 }, 00:15:46.281 { 00:15:46.281 "name": "BaseBdev4", 00:15:46.281 "uuid": "2f968fda-716a-4d74-a09c-141e9728493b", 00:15:46.281 "is_configured": true, 00:15:46.281 "data_offset": 0, 00:15:46.281 "data_size": 65536 00:15:46.281 } 00:15:46.281 ] 00:15:46.281 }' 00:15:46.281 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:46.281 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.540 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.540 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:15:46.540 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.540 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.540 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.800 [2024-11-20 16:03:44.797440] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.800 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:46.800 "name": "Existed_Raid", 00:15:46.800 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:46.800 "strip_size_kb": 64, 00:15:46.800 "state": "configuring", 00:15:46.800 "raid_level": "raid5f", 00:15:46.800 "superblock": false, 00:15:46.800 "num_base_bdevs": 4, 00:15:46.800 "num_base_bdevs_discovered": 2, 00:15:46.800 "num_base_bdevs_operational": 4, 00:15:46.800 "base_bdevs_list": [ 00:15:46.800 { 00:15:46.800 "name": null, 00:15:46.800 "uuid": "9cb206a8-bb09-4f66-9b2c-04a931ac5437", 00:15:46.800 "is_configured": false, 00:15:46.800 "data_offset": 0, 00:15:46.800 "data_size": 65536 00:15:46.800 }, 00:15:46.800 { 00:15:46.800 "name": null, 00:15:46.800 "uuid": "26e7d4b2-a0d4-4d7e-957e-26073e3387bc", 00:15:46.800 "is_configured": false, 00:15:46.800 "data_offset": 0, 00:15:46.800 "data_size": 65536 00:15:46.800 }, 00:15:46.800 { 00:15:46.800 "name": "BaseBdev3", 00:15:46.800 "uuid": "c7789c72-78b4-42dd-8ea7-ecaa7a28a5d8", 00:15:46.800 "is_configured": true, 00:15:46.801 "data_offset": 0, 00:15:46.801 "data_size": 65536 00:15:46.801 }, 00:15:46.801 { 00:15:46.801 "name": "BaseBdev4", 00:15:46.801 "uuid": "2f968fda-716a-4d74-a09c-141e9728493b", 00:15:46.801 "is_configured": true, 00:15:46.801 "data_offset": 0, 00:15:46.801 "data_size": 65536 00:15:46.801 } 00:15:46.801 ] 00:15:46.801 }' 00:15:46.801 16:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:46.801 16:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:47.061 [2024-11-20 16:03:45.196774] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:47.061 "name": "Existed_Raid", 00:15:47.061 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:47.061 "strip_size_kb": 64, 00:15:47.061 "state": "configuring", 00:15:47.061 "raid_level": "raid5f", 00:15:47.061 "superblock": false, 00:15:47.061 "num_base_bdevs": 4, 00:15:47.061 "num_base_bdevs_discovered": 3, 00:15:47.061 "num_base_bdevs_operational": 4, 00:15:47.061 "base_bdevs_list": [ 00:15:47.061 { 00:15:47.061 "name": null, 00:15:47.061 "uuid": "9cb206a8-bb09-4f66-9b2c-04a931ac5437", 00:15:47.061 "is_configured": false, 00:15:47.061 "data_offset": 0, 00:15:47.061 "data_size": 65536 00:15:47.061 }, 00:15:47.061 { 00:15:47.061 "name": "BaseBdev2", 00:15:47.061 "uuid": "26e7d4b2-a0d4-4d7e-957e-26073e3387bc", 00:15:47.061 "is_configured": true, 00:15:47.061 "data_offset": 0, 00:15:47.061 "data_size": 65536 00:15:47.061 }, 00:15:47.061 { 00:15:47.061 "name": "BaseBdev3", 00:15:47.061 "uuid": "c7789c72-78b4-42dd-8ea7-ecaa7a28a5d8", 00:15:47.061 "is_configured": true, 00:15:47.061 "data_offset": 0, 00:15:47.061 "data_size": 65536 00:15:47.061 }, 00:15:47.061 { 00:15:47.061 "name": "BaseBdev4", 00:15:47.061 "uuid": "2f968fda-716a-4d74-a09c-141e9728493b", 00:15:47.061 "is_configured": true, 00:15:47.061 "data_offset": 0, 00:15:47.061 "data_size": 65536 00:15:47.061 } 00:15:47.061 ] 00:15:47.061 }' 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:47.061 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:47.322 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:47.322 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:15:47.322 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.322 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:47.322 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.322 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:15:47.322 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:47.322 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.322 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:15:47.322 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 9cb206a8-bb09-4f66-9b2c-04a931ac5437 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:47.585 [2024-11-20 16:03:45.615743] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:15:47.585 [2024-11-20 16:03:45.615788] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:15:47.585 [2024-11-20 16:03:45.615795] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:15:47.585 [2024-11-20 16:03:45.616046] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:15:47.585 [2024-11-20 16:03:45.620791] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:15:47.585 [2024-11-20 16:03:45.620823] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:15:47.585 [2024-11-20 16:03:45.621042] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:47.585 NewBaseBdev 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:47.585 [ 00:15:47.585 { 00:15:47.585 "name": "NewBaseBdev", 00:15:47.585 "aliases": [ 00:15:47.585 "9cb206a8-bb09-4f66-9b2c-04a931ac5437" 00:15:47.585 ], 00:15:47.585 "product_name": "Malloc disk", 00:15:47.585 "block_size": 512, 00:15:47.585 "num_blocks": 65536, 00:15:47.585 "uuid": "9cb206a8-bb09-4f66-9b2c-04a931ac5437", 00:15:47.585 "assigned_rate_limits": { 00:15:47.585 "rw_ios_per_sec": 0, 00:15:47.585 "rw_mbytes_per_sec": 0, 00:15:47.585 "r_mbytes_per_sec": 0, 00:15:47.585 "w_mbytes_per_sec": 0 00:15:47.585 }, 00:15:47.585 "claimed": true, 00:15:47.585 "claim_type": "exclusive_write", 00:15:47.585 "zoned": false, 00:15:47.585 "supported_io_types": { 00:15:47.585 "read": true, 00:15:47.585 "write": true, 00:15:47.585 "unmap": true, 00:15:47.585 "flush": true, 00:15:47.585 "reset": true, 00:15:47.585 "nvme_admin": false, 00:15:47.585 "nvme_io": false, 00:15:47.585 "nvme_io_md": false, 00:15:47.585 "write_zeroes": true, 00:15:47.585 "zcopy": true, 00:15:47.585 "get_zone_info": false, 00:15:47.585 "zone_management": false, 00:15:47.585 "zone_append": false, 00:15:47.585 "compare": false, 00:15:47.585 "compare_and_write": false, 00:15:47.585 "abort": true, 00:15:47.585 "seek_hole": false, 00:15:47.585 "seek_data": false, 00:15:47.585 "copy": true, 00:15:47.585 "nvme_iov_md": false 00:15:47.585 }, 00:15:47.585 "memory_domains": [ 00:15:47.585 { 00:15:47.585 "dma_device_id": "system", 00:15:47.585 "dma_device_type": 1 00:15:47.585 }, 00:15:47.585 { 00:15:47.585 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:47.585 "dma_device_type": 2 00:15:47.585 } 00:15:47.585 ], 00:15:47.585 "driver_specific": {} 00:15:47.585 } 00:15:47.585 ] 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.585 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:47.585 "name": "Existed_Raid", 00:15:47.585 "uuid": "4495236c-f69f-4a9d-9016-a21da74c1d5f", 00:15:47.585 "strip_size_kb": 64, 00:15:47.585 "state": "online", 00:15:47.585 "raid_level": "raid5f", 00:15:47.585 "superblock": false, 00:15:47.585 "num_base_bdevs": 4, 00:15:47.585 "num_base_bdevs_discovered": 4, 00:15:47.585 "num_base_bdevs_operational": 4, 00:15:47.585 "base_bdevs_list": [ 00:15:47.585 { 00:15:47.585 "name": "NewBaseBdev", 00:15:47.585 "uuid": "9cb206a8-bb09-4f66-9b2c-04a931ac5437", 00:15:47.585 "is_configured": true, 00:15:47.585 "data_offset": 0, 00:15:47.585 "data_size": 65536 00:15:47.585 }, 00:15:47.585 { 00:15:47.585 "name": "BaseBdev2", 00:15:47.585 "uuid": "26e7d4b2-a0d4-4d7e-957e-26073e3387bc", 00:15:47.585 "is_configured": true, 00:15:47.585 "data_offset": 0, 00:15:47.586 "data_size": 65536 00:15:47.586 }, 00:15:47.586 { 00:15:47.586 "name": "BaseBdev3", 00:15:47.586 "uuid": "c7789c72-78b4-42dd-8ea7-ecaa7a28a5d8", 00:15:47.586 "is_configured": true, 00:15:47.586 "data_offset": 0, 00:15:47.586 "data_size": 65536 00:15:47.586 }, 00:15:47.586 { 00:15:47.586 "name": "BaseBdev4", 00:15:47.586 "uuid": "2f968fda-716a-4d74-a09c-141e9728493b", 00:15:47.586 "is_configured": true, 00:15:47.586 "data_offset": 0, 00:15:47.586 "data_size": 65536 00:15:47.586 } 00:15:47.586 ] 00:15:47.586 }' 00:15:47.586 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:47.586 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:47.847 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:15:47.847 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:47.847 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:47.847 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:47.847 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:15:47.847 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:47.847 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:47.847 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:47.847 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.847 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:47.847 [2024-11-20 16:03:45.954607] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:47.847 16:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.847 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:47.847 "name": "Existed_Raid", 00:15:47.847 "aliases": [ 00:15:47.847 "4495236c-f69f-4a9d-9016-a21da74c1d5f" 00:15:47.847 ], 00:15:47.847 "product_name": "Raid Volume", 00:15:47.847 "block_size": 512, 00:15:47.847 "num_blocks": 196608, 00:15:47.847 "uuid": "4495236c-f69f-4a9d-9016-a21da74c1d5f", 00:15:47.847 "assigned_rate_limits": { 00:15:47.847 "rw_ios_per_sec": 0, 00:15:47.847 "rw_mbytes_per_sec": 0, 00:15:47.847 "r_mbytes_per_sec": 0, 00:15:47.847 "w_mbytes_per_sec": 0 00:15:47.847 }, 00:15:47.847 "claimed": false, 00:15:47.847 "zoned": false, 00:15:47.847 "supported_io_types": { 00:15:47.847 "read": true, 00:15:47.847 "write": true, 00:15:47.847 "unmap": false, 00:15:47.847 "flush": false, 00:15:47.847 "reset": true, 00:15:47.847 "nvme_admin": false, 00:15:47.847 "nvme_io": false, 00:15:47.847 "nvme_io_md": false, 00:15:47.847 "write_zeroes": true, 00:15:47.847 "zcopy": false, 00:15:47.847 "get_zone_info": false, 00:15:47.847 "zone_management": false, 00:15:47.847 "zone_append": false, 00:15:47.847 "compare": false, 00:15:47.847 "compare_and_write": false, 00:15:47.847 "abort": false, 00:15:47.847 "seek_hole": false, 00:15:47.847 "seek_data": false, 00:15:47.847 "copy": false, 00:15:47.847 "nvme_iov_md": false 00:15:47.847 }, 00:15:47.847 "driver_specific": { 00:15:47.847 "raid": { 00:15:47.847 "uuid": "4495236c-f69f-4a9d-9016-a21da74c1d5f", 00:15:47.847 "strip_size_kb": 64, 00:15:47.847 "state": "online", 00:15:47.847 "raid_level": "raid5f", 00:15:47.847 "superblock": false, 00:15:47.847 "num_base_bdevs": 4, 00:15:47.847 "num_base_bdevs_discovered": 4, 00:15:47.847 "num_base_bdevs_operational": 4, 00:15:47.847 "base_bdevs_list": [ 00:15:47.847 { 00:15:47.847 "name": "NewBaseBdev", 00:15:47.847 "uuid": "9cb206a8-bb09-4f66-9b2c-04a931ac5437", 00:15:47.847 "is_configured": true, 00:15:47.847 "data_offset": 0, 00:15:47.847 "data_size": 65536 00:15:47.847 }, 00:15:47.847 { 00:15:47.847 "name": "BaseBdev2", 00:15:47.847 "uuid": "26e7d4b2-a0d4-4d7e-957e-26073e3387bc", 00:15:47.847 "is_configured": true, 00:15:47.847 "data_offset": 0, 00:15:47.847 "data_size": 65536 00:15:47.847 }, 00:15:47.847 { 00:15:47.847 "name": "BaseBdev3", 00:15:47.847 "uuid": "c7789c72-78b4-42dd-8ea7-ecaa7a28a5d8", 00:15:47.847 "is_configured": true, 00:15:47.847 "data_offset": 0, 00:15:47.847 "data_size": 65536 00:15:47.847 }, 00:15:47.847 { 00:15:47.847 "name": "BaseBdev4", 00:15:47.847 "uuid": "2f968fda-716a-4d74-a09c-141e9728493b", 00:15:47.847 "is_configured": true, 00:15:47.847 "data_offset": 0, 00:15:47.847 "data_size": 65536 00:15:47.847 } 00:15:47.847 ] 00:15:47.847 } 00:15:47.847 } 00:15:47.847 }' 00:15:47.847 16:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:47.847 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:15:47.847 BaseBdev2 00:15:47.847 BaseBdev3 00:15:47.847 BaseBdev4' 00:15:47.847 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:47.847 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:47.847 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:47.847 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:47.847 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:15:47.847 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.847 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:47.847 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.847 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:47.848 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:47.848 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:47.848 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:47.848 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:47.848 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.848 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:48.108 [2024-11-20 16:03:46.190415] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:48.108 [2024-11-20 16:03:46.190534] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:48.108 [2024-11-20 16:03:46.190650] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:48.108 [2024-11-20 16:03:46.190975] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:48.108 [2024-11-20 16:03:46.191053] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 80479 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 80479 ']' 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # kill -0 80479 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # uname 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 80479 00:15:48.108 killing process with pid 80479 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 80479' 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@973 -- # kill 80479 00:15:48.108 [2024-11-20 16:03:46.223059] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:48.108 16:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@978 -- # wait 80479 00:15:48.369 [2024-11-20 16:03:46.470074] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:48.937 ************************************ 00:15:48.937 END TEST raid5f_state_function_test 00:15:48.937 ************************************ 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:15:48.937 00:15:48.937 real 0m8.390s 00:15:48.937 user 0m13.429s 00:15:48.937 sys 0m1.337s 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:48.937 16:03:47 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 4 true 00:15:48.937 16:03:47 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:15:48.937 16:03:47 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:48.937 16:03:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:48.937 ************************************ 00:15:48.937 START TEST raid5f_state_function_test_sb 00:15:48.937 ************************************ 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 4 true 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=81123 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:15:48.937 Process raid pid: 81123 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 81123' 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 81123 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 81123 ']' 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:48.937 16:03:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:48.938 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:48.938 16:03:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:48.938 16:03:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:49.201 [2024-11-20 16:03:47.235999] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:15:49.201 [2024-11-20 16:03:47.236276] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:49.201 [2024-11-20 16:03:47.392439] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:49.463 [2024-11-20 16:03:47.504354] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:49.463 [2024-11-20 16:03:47.659760] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:49.463 [2024-11-20 16:03:47.659801] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.036 [2024-11-20 16:03:48.112864] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:50.036 [2024-11-20 16:03:48.112924] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:50.036 [2024-11-20 16:03:48.112935] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:50.036 [2024-11-20 16:03:48.112944] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:50.036 [2024-11-20 16:03:48.112951] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:50.036 [2024-11-20 16:03:48.112959] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:50.036 [2024-11-20 16:03:48.112965] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:15:50.036 [2024-11-20 16:03:48.112973] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:50.036 "name": "Existed_Raid", 00:15:50.036 "uuid": "0b136111-b97a-4276-b500-c5a1a939dd13", 00:15:50.036 "strip_size_kb": 64, 00:15:50.036 "state": "configuring", 00:15:50.036 "raid_level": "raid5f", 00:15:50.036 "superblock": true, 00:15:50.036 "num_base_bdevs": 4, 00:15:50.036 "num_base_bdevs_discovered": 0, 00:15:50.036 "num_base_bdevs_operational": 4, 00:15:50.036 "base_bdevs_list": [ 00:15:50.036 { 00:15:50.036 "name": "BaseBdev1", 00:15:50.036 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:50.036 "is_configured": false, 00:15:50.036 "data_offset": 0, 00:15:50.036 "data_size": 0 00:15:50.036 }, 00:15:50.036 { 00:15:50.036 "name": "BaseBdev2", 00:15:50.036 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:50.036 "is_configured": false, 00:15:50.036 "data_offset": 0, 00:15:50.036 "data_size": 0 00:15:50.036 }, 00:15:50.036 { 00:15:50.036 "name": "BaseBdev3", 00:15:50.036 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:50.036 "is_configured": false, 00:15:50.036 "data_offset": 0, 00:15:50.036 "data_size": 0 00:15:50.036 }, 00:15:50.036 { 00:15:50.036 "name": "BaseBdev4", 00:15:50.036 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:50.036 "is_configured": false, 00:15:50.036 "data_offset": 0, 00:15:50.036 "data_size": 0 00:15:50.036 } 00:15:50.036 ] 00:15:50.036 }' 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:50.036 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.297 [2024-11-20 16:03:48.448876] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:50.297 [2024-11-20 16:03:48.448913] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.297 [2024-11-20 16:03:48.456893] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:50.297 [2024-11-20 16:03:48.456934] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:50.297 [2024-11-20 16:03:48.456942] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:50.297 [2024-11-20 16:03:48.456952] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:50.297 [2024-11-20 16:03:48.456958] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:50.297 [2024-11-20 16:03:48.456967] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:50.297 [2024-11-20 16:03:48.456973] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:15:50.297 [2024-11-20 16:03:48.456982] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.297 [2024-11-20 16:03:48.489311] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:50.297 BaseBdev1 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.297 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.297 [ 00:15:50.297 { 00:15:50.297 "name": "BaseBdev1", 00:15:50.297 "aliases": [ 00:15:50.297 "ea827053-b805-47b4-92fc-a151903a314e" 00:15:50.297 ], 00:15:50.297 "product_name": "Malloc disk", 00:15:50.297 "block_size": 512, 00:15:50.297 "num_blocks": 65536, 00:15:50.297 "uuid": "ea827053-b805-47b4-92fc-a151903a314e", 00:15:50.297 "assigned_rate_limits": { 00:15:50.297 "rw_ios_per_sec": 0, 00:15:50.297 "rw_mbytes_per_sec": 0, 00:15:50.297 "r_mbytes_per_sec": 0, 00:15:50.297 "w_mbytes_per_sec": 0 00:15:50.297 }, 00:15:50.297 "claimed": true, 00:15:50.297 "claim_type": "exclusive_write", 00:15:50.297 "zoned": false, 00:15:50.297 "supported_io_types": { 00:15:50.297 "read": true, 00:15:50.297 "write": true, 00:15:50.297 "unmap": true, 00:15:50.297 "flush": true, 00:15:50.297 "reset": true, 00:15:50.297 "nvme_admin": false, 00:15:50.297 "nvme_io": false, 00:15:50.297 "nvme_io_md": false, 00:15:50.297 "write_zeroes": true, 00:15:50.297 "zcopy": true, 00:15:50.297 "get_zone_info": false, 00:15:50.297 "zone_management": false, 00:15:50.297 "zone_append": false, 00:15:50.297 "compare": false, 00:15:50.297 "compare_and_write": false, 00:15:50.297 "abort": true, 00:15:50.297 "seek_hole": false, 00:15:50.297 "seek_data": false, 00:15:50.297 "copy": true, 00:15:50.297 "nvme_iov_md": false 00:15:50.297 }, 00:15:50.297 "memory_domains": [ 00:15:50.297 { 00:15:50.297 "dma_device_id": "system", 00:15:50.297 "dma_device_type": 1 00:15:50.297 }, 00:15:50.297 { 00:15:50.297 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:50.297 "dma_device_type": 2 00:15:50.297 } 00:15:50.297 ], 00:15:50.297 "driver_specific": {} 00:15:50.297 } 00:15:50.297 ] 00:15:50.298 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.298 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:50.298 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:50.298 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:50.298 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:50.298 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:50.298 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:50.298 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:50.298 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:50.298 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:50.298 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:50.298 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:50.298 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.298 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:50.298 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.298 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.298 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.562 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:50.562 "name": "Existed_Raid", 00:15:50.562 "uuid": "2c239a7b-f1de-46eb-bfb1-1e85ae3b5828", 00:15:50.562 "strip_size_kb": 64, 00:15:50.562 "state": "configuring", 00:15:50.562 "raid_level": "raid5f", 00:15:50.562 "superblock": true, 00:15:50.562 "num_base_bdevs": 4, 00:15:50.562 "num_base_bdevs_discovered": 1, 00:15:50.562 "num_base_bdevs_operational": 4, 00:15:50.562 "base_bdevs_list": [ 00:15:50.562 { 00:15:50.562 "name": "BaseBdev1", 00:15:50.562 "uuid": "ea827053-b805-47b4-92fc-a151903a314e", 00:15:50.562 "is_configured": true, 00:15:50.562 "data_offset": 2048, 00:15:50.562 "data_size": 63488 00:15:50.562 }, 00:15:50.562 { 00:15:50.562 "name": "BaseBdev2", 00:15:50.562 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:50.562 "is_configured": false, 00:15:50.562 "data_offset": 0, 00:15:50.562 "data_size": 0 00:15:50.562 }, 00:15:50.562 { 00:15:50.562 "name": "BaseBdev3", 00:15:50.563 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:50.563 "is_configured": false, 00:15:50.563 "data_offset": 0, 00:15:50.563 "data_size": 0 00:15:50.563 }, 00:15:50.563 { 00:15:50.563 "name": "BaseBdev4", 00:15:50.563 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:50.563 "is_configured": false, 00:15:50.563 "data_offset": 0, 00:15:50.563 "data_size": 0 00:15:50.563 } 00:15:50.563 ] 00:15:50.563 }' 00:15:50.563 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:50.563 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.821 [2024-11-20 16:03:48.861434] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:50.821 [2024-11-20 16:03:48.861620] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.821 [2024-11-20 16:03:48.869501] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:50.821 [2024-11-20 16:03:48.871424] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:50.821 [2024-11-20 16:03:48.871541] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:50.821 [2024-11-20 16:03:48.871600] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:50.821 [2024-11-20 16:03:48.871629] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:50.821 [2024-11-20 16:03:48.871694] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:15:50.821 [2024-11-20 16:03:48.871721] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:50.821 "name": "Existed_Raid", 00:15:50.821 "uuid": "eff77c23-c196-4dfc-9773-1b1dd824e980", 00:15:50.821 "strip_size_kb": 64, 00:15:50.821 "state": "configuring", 00:15:50.821 "raid_level": "raid5f", 00:15:50.821 "superblock": true, 00:15:50.821 "num_base_bdevs": 4, 00:15:50.821 "num_base_bdevs_discovered": 1, 00:15:50.821 "num_base_bdevs_operational": 4, 00:15:50.821 "base_bdevs_list": [ 00:15:50.821 { 00:15:50.821 "name": "BaseBdev1", 00:15:50.821 "uuid": "ea827053-b805-47b4-92fc-a151903a314e", 00:15:50.821 "is_configured": true, 00:15:50.821 "data_offset": 2048, 00:15:50.821 "data_size": 63488 00:15:50.821 }, 00:15:50.821 { 00:15:50.821 "name": "BaseBdev2", 00:15:50.821 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:50.821 "is_configured": false, 00:15:50.821 "data_offset": 0, 00:15:50.821 "data_size": 0 00:15:50.821 }, 00:15:50.821 { 00:15:50.821 "name": "BaseBdev3", 00:15:50.821 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:50.821 "is_configured": false, 00:15:50.821 "data_offset": 0, 00:15:50.821 "data_size": 0 00:15:50.821 }, 00:15:50.821 { 00:15:50.821 "name": "BaseBdev4", 00:15:50.821 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:50.821 "is_configured": false, 00:15:50.821 "data_offset": 0, 00:15:50.821 "data_size": 0 00:15:50.821 } 00:15:50.821 ] 00:15:50.821 }' 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:50.821 16:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.081 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:15:51.081 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.081 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.081 [2024-11-20 16:03:49.216107] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:51.081 BaseBdev2 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.082 [ 00:15:51.082 { 00:15:51.082 "name": "BaseBdev2", 00:15:51.082 "aliases": [ 00:15:51.082 "06bc719e-5949-4bcd-a128-494ba28e5170" 00:15:51.082 ], 00:15:51.082 "product_name": "Malloc disk", 00:15:51.082 "block_size": 512, 00:15:51.082 "num_blocks": 65536, 00:15:51.082 "uuid": "06bc719e-5949-4bcd-a128-494ba28e5170", 00:15:51.082 "assigned_rate_limits": { 00:15:51.082 "rw_ios_per_sec": 0, 00:15:51.082 "rw_mbytes_per_sec": 0, 00:15:51.082 "r_mbytes_per_sec": 0, 00:15:51.082 "w_mbytes_per_sec": 0 00:15:51.082 }, 00:15:51.082 "claimed": true, 00:15:51.082 "claim_type": "exclusive_write", 00:15:51.082 "zoned": false, 00:15:51.082 "supported_io_types": { 00:15:51.082 "read": true, 00:15:51.082 "write": true, 00:15:51.082 "unmap": true, 00:15:51.082 "flush": true, 00:15:51.082 "reset": true, 00:15:51.082 "nvme_admin": false, 00:15:51.082 "nvme_io": false, 00:15:51.082 "nvme_io_md": false, 00:15:51.082 "write_zeroes": true, 00:15:51.082 "zcopy": true, 00:15:51.082 "get_zone_info": false, 00:15:51.082 "zone_management": false, 00:15:51.082 "zone_append": false, 00:15:51.082 "compare": false, 00:15:51.082 "compare_and_write": false, 00:15:51.082 "abort": true, 00:15:51.082 "seek_hole": false, 00:15:51.082 "seek_data": false, 00:15:51.082 "copy": true, 00:15:51.082 "nvme_iov_md": false 00:15:51.082 }, 00:15:51.082 "memory_domains": [ 00:15:51.082 { 00:15:51.082 "dma_device_id": "system", 00:15:51.082 "dma_device_type": 1 00:15:51.082 }, 00:15:51.082 { 00:15:51.082 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:51.082 "dma_device_type": 2 00:15:51.082 } 00:15:51.082 ], 00:15:51.082 "driver_specific": {} 00:15:51.082 } 00:15:51.082 ] 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:51.082 "name": "Existed_Raid", 00:15:51.082 "uuid": "eff77c23-c196-4dfc-9773-1b1dd824e980", 00:15:51.082 "strip_size_kb": 64, 00:15:51.082 "state": "configuring", 00:15:51.082 "raid_level": "raid5f", 00:15:51.082 "superblock": true, 00:15:51.082 "num_base_bdevs": 4, 00:15:51.082 "num_base_bdevs_discovered": 2, 00:15:51.082 "num_base_bdevs_operational": 4, 00:15:51.082 "base_bdevs_list": [ 00:15:51.082 { 00:15:51.082 "name": "BaseBdev1", 00:15:51.082 "uuid": "ea827053-b805-47b4-92fc-a151903a314e", 00:15:51.082 "is_configured": true, 00:15:51.082 "data_offset": 2048, 00:15:51.082 "data_size": 63488 00:15:51.082 }, 00:15:51.082 { 00:15:51.082 "name": "BaseBdev2", 00:15:51.082 "uuid": "06bc719e-5949-4bcd-a128-494ba28e5170", 00:15:51.082 "is_configured": true, 00:15:51.082 "data_offset": 2048, 00:15:51.082 "data_size": 63488 00:15:51.082 }, 00:15:51.082 { 00:15:51.082 "name": "BaseBdev3", 00:15:51.082 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:51.082 "is_configured": false, 00:15:51.082 "data_offset": 0, 00:15:51.082 "data_size": 0 00:15:51.082 }, 00:15:51.082 { 00:15:51.082 "name": "BaseBdev4", 00:15:51.082 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:51.082 "is_configured": false, 00:15:51.082 "data_offset": 0, 00:15:51.082 "data_size": 0 00:15:51.082 } 00:15:51.082 ] 00:15:51.082 }' 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:51.082 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.343 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:15:51.343 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.343 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.603 [2024-11-20 16:03:49.616850] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:51.603 BaseBdev3 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.603 [ 00:15:51.603 { 00:15:51.603 "name": "BaseBdev3", 00:15:51.603 "aliases": [ 00:15:51.603 "0ee8fc52-5e1a-4d83-8b21-fd894ae691d6" 00:15:51.603 ], 00:15:51.603 "product_name": "Malloc disk", 00:15:51.603 "block_size": 512, 00:15:51.603 "num_blocks": 65536, 00:15:51.603 "uuid": "0ee8fc52-5e1a-4d83-8b21-fd894ae691d6", 00:15:51.603 "assigned_rate_limits": { 00:15:51.603 "rw_ios_per_sec": 0, 00:15:51.603 "rw_mbytes_per_sec": 0, 00:15:51.603 "r_mbytes_per_sec": 0, 00:15:51.603 "w_mbytes_per_sec": 0 00:15:51.603 }, 00:15:51.603 "claimed": true, 00:15:51.603 "claim_type": "exclusive_write", 00:15:51.603 "zoned": false, 00:15:51.603 "supported_io_types": { 00:15:51.603 "read": true, 00:15:51.603 "write": true, 00:15:51.603 "unmap": true, 00:15:51.603 "flush": true, 00:15:51.603 "reset": true, 00:15:51.603 "nvme_admin": false, 00:15:51.603 "nvme_io": false, 00:15:51.603 "nvme_io_md": false, 00:15:51.603 "write_zeroes": true, 00:15:51.603 "zcopy": true, 00:15:51.603 "get_zone_info": false, 00:15:51.603 "zone_management": false, 00:15:51.603 "zone_append": false, 00:15:51.603 "compare": false, 00:15:51.603 "compare_and_write": false, 00:15:51.603 "abort": true, 00:15:51.603 "seek_hole": false, 00:15:51.603 "seek_data": false, 00:15:51.603 "copy": true, 00:15:51.603 "nvme_iov_md": false 00:15:51.603 }, 00:15:51.603 "memory_domains": [ 00:15:51.603 { 00:15:51.603 "dma_device_id": "system", 00:15:51.603 "dma_device_type": 1 00:15:51.603 }, 00:15:51.603 { 00:15:51.603 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:51.603 "dma_device_type": 2 00:15:51.603 } 00:15:51.603 ], 00:15:51.603 "driver_specific": {} 00:15:51.603 } 00:15:51.603 ] 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.603 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.604 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.604 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:51.604 "name": "Existed_Raid", 00:15:51.604 "uuid": "eff77c23-c196-4dfc-9773-1b1dd824e980", 00:15:51.604 "strip_size_kb": 64, 00:15:51.604 "state": "configuring", 00:15:51.604 "raid_level": "raid5f", 00:15:51.604 "superblock": true, 00:15:51.604 "num_base_bdevs": 4, 00:15:51.604 "num_base_bdevs_discovered": 3, 00:15:51.604 "num_base_bdevs_operational": 4, 00:15:51.604 "base_bdevs_list": [ 00:15:51.604 { 00:15:51.604 "name": "BaseBdev1", 00:15:51.604 "uuid": "ea827053-b805-47b4-92fc-a151903a314e", 00:15:51.604 "is_configured": true, 00:15:51.604 "data_offset": 2048, 00:15:51.604 "data_size": 63488 00:15:51.604 }, 00:15:51.604 { 00:15:51.604 "name": "BaseBdev2", 00:15:51.604 "uuid": "06bc719e-5949-4bcd-a128-494ba28e5170", 00:15:51.604 "is_configured": true, 00:15:51.604 "data_offset": 2048, 00:15:51.604 "data_size": 63488 00:15:51.604 }, 00:15:51.604 { 00:15:51.604 "name": "BaseBdev3", 00:15:51.604 "uuid": "0ee8fc52-5e1a-4d83-8b21-fd894ae691d6", 00:15:51.604 "is_configured": true, 00:15:51.604 "data_offset": 2048, 00:15:51.604 "data_size": 63488 00:15:51.604 }, 00:15:51.604 { 00:15:51.604 "name": "BaseBdev4", 00:15:51.604 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:51.604 "is_configured": false, 00:15:51.604 "data_offset": 0, 00:15:51.604 "data_size": 0 00:15:51.604 } 00:15:51.604 ] 00:15:51.604 }' 00:15:51.604 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:51.604 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.865 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:15:51.865 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.865 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.865 [2024-11-20 16:03:49.979827] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:51.865 [2024-11-20 16:03:49.980064] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:15:51.865 [2024-11-20 16:03:49.980078] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:51.865 [2024-11-20 16:03:49.980331] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:15:51.865 BaseBdev4 00:15:51.865 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.865 16:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:15:51.865 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:15:51.865 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:51.865 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:51.865 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:51.865 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:51.865 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:51.865 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.865 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.865 [2024-11-20 16:03:49.985352] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:15:51.865 [2024-11-20 16:03:49.985493] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:15:51.865 [2024-11-20 16:03:49.985822] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:51.865 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.865 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:15:51.865 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.865 16:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.865 [ 00:15:51.865 { 00:15:51.865 "name": "BaseBdev4", 00:15:51.865 "aliases": [ 00:15:51.865 "2bbda38b-cf66-478e-bb2b-7c1b2ae044bd" 00:15:51.865 ], 00:15:51.865 "product_name": "Malloc disk", 00:15:51.865 "block_size": 512, 00:15:51.865 "num_blocks": 65536, 00:15:51.865 "uuid": "2bbda38b-cf66-478e-bb2b-7c1b2ae044bd", 00:15:51.865 "assigned_rate_limits": { 00:15:51.865 "rw_ios_per_sec": 0, 00:15:51.865 "rw_mbytes_per_sec": 0, 00:15:51.865 "r_mbytes_per_sec": 0, 00:15:51.865 "w_mbytes_per_sec": 0 00:15:51.865 }, 00:15:51.865 "claimed": true, 00:15:51.865 "claim_type": "exclusive_write", 00:15:51.865 "zoned": false, 00:15:51.865 "supported_io_types": { 00:15:51.865 "read": true, 00:15:51.865 "write": true, 00:15:51.865 "unmap": true, 00:15:51.865 "flush": true, 00:15:51.865 "reset": true, 00:15:51.865 "nvme_admin": false, 00:15:51.865 "nvme_io": false, 00:15:51.865 "nvme_io_md": false, 00:15:51.865 "write_zeroes": true, 00:15:51.865 "zcopy": true, 00:15:51.865 "get_zone_info": false, 00:15:51.865 "zone_management": false, 00:15:51.865 "zone_append": false, 00:15:51.865 "compare": false, 00:15:51.865 "compare_and_write": false, 00:15:51.865 "abort": true, 00:15:51.865 "seek_hole": false, 00:15:51.865 "seek_data": false, 00:15:51.865 "copy": true, 00:15:51.865 "nvme_iov_md": false 00:15:51.865 }, 00:15:51.865 "memory_domains": [ 00:15:51.865 { 00:15:51.865 "dma_device_id": "system", 00:15:51.865 "dma_device_type": 1 00:15:51.865 }, 00:15:51.865 { 00:15:51.865 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:51.865 "dma_device_type": 2 00:15:51.865 } 00:15:51.865 ], 00:15:51.865 "driver_specific": {} 00:15:51.865 } 00:15:51.865 ] 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.865 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:51.865 "name": "Existed_Raid", 00:15:51.865 "uuid": "eff77c23-c196-4dfc-9773-1b1dd824e980", 00:15:51.865 "strip_size_kb": 64, 00:15:51.865 "state": "online", 00:15:51.865 "raid_level": "raid5f", 00:15:51.865 "superblock": true, 00:15:51.865 "num_base_bdevs": 4, 00:15:51.865 "num_base_bdevs_discovered": 4, 00:15:51.865 "num_base_bdevs_operational": 4, 00:15:51.865 "base_bdevs_list": [ 00:15:51.865 { 00:15:51.865 "name": "BaseBdev1", 00:15:51.865 "uuid": "ea827053-b805-47b4-92fc-a151903a314e", 00:15:51.865 "is_configured": true, 00:15:51.865 "data_offset": 2048, 00:15:51.865 "data_size": 63488 00:15:51.865 }, 00:15:51.866 { 00:15:51.866 "name": "BaseBdev2", 00:15:51.866 "uuid": "06bc719e-5949-4bcd-a128-494ba28e5170", 00:15:51.866 "is_configured": true, 00:15:51.866 "data_offset": 2048, 00:15:51.866 "data_size": 63488 00:15:51.866 }, 00:15:51.866 { 00:15:51.866 "name": "BaseBdev3", 00:15:51.866 "uuid": "0ee8fc52-5e1a-4d83-8b21-fd894ae691d6", 00:15:51.866 "is_configured": true, 00:15:51.866 "data_offset": 2048, 00:15:51.866 "data_size": 63488 00:15:51.866 }, 00:15:51.866 { 00:15:51.866 "name": "BaseBdev4", 00:15:51.866 "uuid": "2bbda38b-cf66-478e-bb2b-7c1b2ae044bd", 00:15:51.866 "is_configured": true, 00:15:51.866 "data_offset": 2048, 00:15:51.866 "data_size": 63488 00:15:51.866 } 00:15:51.866 ] 00:15:51.866 }' 00:15:51.866 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:51.866 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.146 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:15:52.146 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:52.146 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:52.146 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:52.146 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:15:52.146 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:52.146 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:52.146 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:52.146 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.146 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.406 [2024-11-20 16:03:50.396552] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:52.406 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.406 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:52.406 "name": "Existed_Raid", 00:15:52.406 "aliases": [ 00:15:52.406 "eff77c23-c196-4dfc-9773-1b1dd824e980" 00:15:52.406 ], 00:15:52.406 "product_name": "Raid Volume", 00:15:52.406 "block_size": 512, 00:15:52.406 "num_blocks": 190464, 00:15:52.406 "uuid": "eff77c23-c196-4dfc-9773-1b1dd824e980", 00:15:52.406 "assigned_rate_limits": { 00:15:52.406 "rw_ios_per_sec": 0, 00:15:52.406 "rw_mbytes_per_sec": 0, 00:15:52.406 "r_mbytes_per_sec": 0, 00:15:52.406 "w_mbytes_per_sec": 0 00:15:52.406 }, 00:15:52.406 "claimed": false, 00:15:52.406 "zoned": false, 00:15:52.406 "supported_io_types": { 00:15:52.406 "read": true, 00:15:52.406 "write": true, 00:15:52.406 "unmap": false, 00:15:52.406 "flush": false, 00:15:52.406 "reset": true, 00:15:52.406 "nvme_admin": false, 00:15:52.406 "nvme_io": false, 00:15:52.406 "nvme_io_md": false, 00:15:52.406 "write_zeroes": true, 00:15:52.406 "zcopy": false, 00:15:52.406 "get_zone_info": false, 00:15:52.406 "zone_management": false, 00:15:52.406 "zone_append": false, 00:15:52.406 "compare": false, 00:15:52.406 "compare_and_write": false, 00:15:52.406 "abort": false, 00:15:52.406 "seek_hole": false, 00:15:52.406 "seek_data": false, 00:15:52.406 "copy": false, 00:15:52.406 "nvme_iov_md": false 00:15:52.406 }, 00:15:52.406 "driver_specific": { 00:15:52.406 "raid": { 00:15:52.406 "uuid": "eff77c23-c196-4dfc-9773-1b1dd824e980", 00:15:52.406 "strip_size_kb": 64, 00:15:52.406 "state": "online", 00:15:52.406 "raid_level": "raid5f", 00:15:52.406 "superblock": true, 00:15:52.406 "num_base_bdevs": 4, 00:15:52.406 "num_base_bdevs_discovered": 4, 00:15:52.406 "num_base_bdevs_operational": 4, 00:15:52.406 "base_bdevs_list": [ 00:15:52.406 { 00:15:52.406 "name": "BaseBdev1", 00:15:52.406 "uuid": "ea827053-b805-47b4-92fc-a151903a314e", 00:15:52.406 "is_configured": true, 00:15:52.406 "data_offset": 2048, 00:15:52.406 "data_size": 63488 00:15:52.406 }, 00:15:52.406 { 00:15:52.406 "name": "BaseBdev2", 00:15:52.406 "uuid": "06bc719e-5949-4bcd-a128-494ba28e5170", 00:15:52.406 "is_configured": true, 00:15:52.406 "data_offset": 2048, 00:15:52.406 "data_size": 63488 00:15:52.406 }, 00:15:52.406 { 00:15:52.406 "name": "BaseBdev3", 00:15:52.406 "uuid": "0ee8fc52-5e1a-4d83-8b21-fd894ae691d6", 00:15:52.406 "is_configured": true, 00:15:52.406 "data_offset": 2048, 00:15:52.406 "data_size": 63488 00:15:52.406 }, 00:15:52.406 { 00:15:52.406 "name": "BaseBdev4", 00:15:52.406 "uuid": "2bbda38b-cf66-478e-bb2b-7c1b2ae044bd", 00:15:52.406 "is_configured": true, 00:15:52.406 "data_offset": 2048, 00:15:52.406 "data_size": 63488 00:15:52.406 } 00:15:52.406 ] 00:15:52.406 } 00:15:52.406 } 00:15:52.406 }' 00:15:52.406 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:52.406 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:15:52.406 BaseBdev2 00:15:52.406 BaseBdev3 00:15:52.406 BaseBdev4' 00:15:52.406 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:52.406 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:52.406 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:52.406 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:15:52.406 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.406 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.407 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.407 [2024-11-20 16:03:50.652448] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:52.667 "name": "Existed_Raid", 00:15:52.667 "uuid": "eff77c23-c196-4dfc-9773-1b1dd824e980", 00:15:52.667 "strip_size_kb": 64, 00:15:52.667 "state": "online", 00:15:52.667 "raid_level": "raid5f", 00:15:52.667 "superblock": true, 00:15:52.667 "num_base_bdevs": 4, 00:15:52.667 "num_base_bdevs_discovered": 3, 00:15:52.667 "num_base_bdevs_operational": 3, 00:15:52.667 "base_bdevs_list": [ 00:15:52.667 { 00:15:52.667 "name": null, 00:15:52.667 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:52.667 "is_configured": false, 00:15:52.667 "data_offset": 0, 00:15:52.667 "data_size": 63488 00:15:52.667 }, 00:15:52.667 { 00:15:52.667 "name": "BaseBdev2", 00:15:52.667 "uuid": "06bc719e-5949-4bcd-a128-494ba28e5170", 00:15:52.667 "is_configured": true, 00:15:52.667 "data_offset": 2048, 00:15:52.667 "data_size": 63488 00:15:52.667 }, 00:15:52.667 { 00:15:52.667 "name": "BaseBdev3", 00:15:52.667 "uuid": "0ee8fc52-5e1a-4d83-8b21-fd894ae691d6", 00:15:52.667 "is_configured": true, 00:15:52.667 "data_offset": 2048, 00:15:52.667 "data_size": 63488 00:15:52.667 }, 00:15:52.667 { 00:15:52.667 "name": "BaseBdev4", 00:15:52.667 "uuid": "2bbda38b-cf66-478e-bb2b-7c1b2ae044bd", 00:15:52.667 "is_configured": true, 00:15:52.667 "data_offset": 2048, 00:15:52.667 "data_size": 63488 00:15:52.667 } 00:15:52.667 ] 00:15:52.667 }' 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:52.667 16:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.926 [2024-11-20 16:03:51.063380] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:52.926 [2024-11-20 16:03:51.063530] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:52.926 [2024-11-20 16:03:51.122598] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.926 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.926 [2024-11-20 16:03:51.162646] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.185 [2024-11-20 16:03:51.259966] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:15:53.185 [2024-11-20 16:03:51.260010] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.185 BaseBdev2 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.185 [ 00:15:53.185 { 00:15:53.185 "name": "BaseBdev2", 00:15:53.185 "aliases": [ 00:15:53.185 "0e549228-823d-407b-bdf3-6decab2c44f2" 00:15:53.185 ], 00:15:53.185 "product_name": "Malloc disk", 00:15:53.185 "block_size": 512, 00:15:53.185 "num_blocks": 65536, 00:15:53.185 "uuid": "0e549228-823d-407b-bdf3-6decab2c44f2", 00:15:53.185 "assigned_rate_limits": { 00:15:53.185 "rw_ios_per_sec": 0, 00:15:53.185 "rw_mbytes_per_sec": 0, 00:15:53.185 "r_mbytes_per_sec": 0, 00:15:53.185 "w_mbytes_per_sec": 0 00:15:53.185 }, 00:15:53.185 "claimed": false, 00:15:53.185 "zoned": false, 00:15:53.185 "supported_io_types": { 00:15:53.185 "read": true, 00:15:53.185 "write": true, 00:15:53.185 "unmap": true, 00:15:53.185 "flush": true, 00:15:53.185 "reset": true, 00:15:53.185 "nvme_admin": false, 00:15:53.185 "nvme_io": false, 00:15:53.185 "nvme_io_md": false, 00:15:53.185 "write_zeroes": true, 00:15:53.185 "zcopy": true, 00:15:53.185 "get_zone_info": false, 00:15:53.185 "zone_management": false, 00:15:53.185 "zone_append": false, 00:15:53.185 "compare": false, 00:15:53.185 "compare_and_write": false, 00:15:53.185 "abort": true, 00:15:53.185 "seek_hole": false, 00:15:53.185 "seek_data": false, 00:15:53.185 "copy": true, 00:15:53.185 "nvme_iov_md": false 00:15:53.185 }, 00:15:53.185 "memory_domains": [ 00:15:53.185 { 00:15:53.185 "dma_device_id": "system", 00:15:53.185 "dma_device_type": 1 00:15:53.185 }, 00:15:53.185 { 00:15:53.185 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:53.185 "dma_device_type": 2 00:15:53.185 } 00:15:53.185 ], 00:15:53.185 "driver_specific": {} 00:15:53.185 } 00:15:53.185 ] 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.185 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.444 BaseBdev3 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.444 [ 00:15:53.444 { 00:15:53.444 "name": "BaseBdev3", 00:15:53.444 "aliases": [ 00:15:53.444 "82beca06-afc9-413a-acae-d2668507b354" 00:15:53.444 ], 00:15:53.444 "product_name": "Malloc disk", 00:15:53.444 "block_size": 512, 00:15:53.444 "num_blocks": 65536, 00:15:53.444 "uuid": "82beca06-afc9-413a-acae-d2668507b354", 00:15:53.444 "assigned_rate_limits": { 00:15:53.444 "rw_ios_per_sec": 0, 00:15:53.444 "rw_mbytes_per_sec": 0, 00:15:53.444 "r_mbytes_per_sec": 0, 00:15:53.444 "w_mbytes_per_sec": 0 00:15:53.444 }, 00:15:53.444 "claimed": false, 00:15:53.444 "zoned": false, 00:15:53.444 "supported_io_types": { 00:15:53.444 "read": true, 00:15:53.444 "write": true, 00:15:53.444 "unmap": true, 00:15:53.444 "flush": true, 00:15:53.444 "reset": true, 00:15:53.444 "nvme_admin": false, 00:15:53.444 "nvme_io": false, 00:15:53.444 "nvme_io_md": false, 00:15:53.444 "write_zeroes": true, 00:15:53.444 "zcopy": true, 00:15:53.444 "get_zone_info": false, 00:15:53.444 "zone_management": false, 00:15:53.444 "zone_append": false, 00:15:53.444 "compare": false, 00:15:53.444 "compare_and_write": false, 00:15:53.444 "abort": true, 00:15:53.444 "seek_hole": false, 00:15:53.444 "seek_data": false, 00:15:53.444 "copy": true, 00:15:53.444 "nvme_iov_md": false 00:15:53.444 }, 00:15:53.444 "memory_domains": [ 00:15:53.444 { 00:15:53.444 "dma_device_id": "system", 00:15:53.444 "dma_device_type": 1 00:15:53.444 }, 00:15:53.444 { 00:15:53.444 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:53.444 "dma_device_type": 2 00:15:53.444 } 00:15:53.444 ], 00:15:53.444 "driver_specific": {} 00:15:53.444 } 00:15:53.444 ] 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.444 BaseBdev4 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.444 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.444 [ 00:15:53.444 { 00:15:53.444 "name": "BaseBdev4", 00:15:53.444 "aliases": [ 00:15:53.444 "1be04fbd-f461-4c95-8552-8938cfcd96b0" 00:15:53.444 ], 00:15:53.444 "product_name": "Malloc disk", 00:15:53.444 "block_size": 512, 00:15:53.444 "num_blocks": 65536, 00:15:53.444 "uuid": "1be04fbd-f461-4c95-8552-8938cfcd96b0", 00:15:53.444 "assigned_rate_limits": { 00:15:53.444 "rw_ios_per_sec": 0, 00:15:53.444 "rw_mbytes_per_sec": 0, 00:15:53.444 "r_mbytes_per_sec": 0, 00:15:53.444 "w_mbytes_per_sec": 0 00:15:53.444 }, 00:15:53.444 "claimed": false, 00:15:53.444 "zoned": false, 00:15:53.444 "supported_io_types": { 00:15:53.444 "read": true, 00:15:53.444 "write": true, 00:15:53.444 "unmap": true, 00:15:53.444 "flush": true, 00:15:53.444 "reset": true, 00:15:53.444 "nvme_admin": false, 00:15:53.444 "nvme_io": false, 00:15:53.444 "nvme_io_md": false, 00:15:53.444 "write_zeroes": true, 00:15:53.444 "zcopy": true, 00:15:53.444 "get_zone_info": false, 00:15:53.444 "zone_management": false, 00:15:53.444 "zone_append": false, 00:15:53.444 "compare": false, 00:15:53.444 "compare_and_write": false, 00:15:53.444 "abort": true, 00:15:53.444 "seek_hole": false, 00:15:53.444 "seek_data": false, 00:15:53.444 "copy": true, 00:15:53.444 "nvme_iov_md": false 00:15:53.444 }, 00:15:53.444 "memory_domains": [ 00:15:53.444 { 00:15:53.444 "dma_device_id": "system", 00:15:53.445 "dma_device_type": 1 00:15:53.445 }, 00:15:53.445 { 00:15:53.445 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:53.445 "dma_device_type": 2 00:15:53.445 } 00:15:53.445 ], 00:15:53.445 "driver_specific": {} 00:15:53.445 } 00:15:53.445 ] 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.445 [2024-11-20 16:03:51.532181] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:53.445 [2024-11-20 16:03:51.532331] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:53.445 [2024-11-20 16:03:51.532406] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:53.445 [2024-11-20 16:03:51.534254] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:53.445 [2024-11-20 16:03:51.534386] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:53.445 "name": "Existed_Raid", 00:15:53.445 "uuid": "3a3914f7-726f-42c3-862a-d21b1fe7fb0e", 00:15:53.445 "strip_size_kb": 64, 00:15:53.445 "state": "configuring", 00:15:53.445 "raid_level": "raid5f", 00:15:53.445 "superblock": true, 00:15:53.445 "num_base_bdevs": 4, 00:15:53.445 "num_base_bdevs_discovered": 3, 00:15:53.445 "num_base_bdevs_operational": 4, 00:15:53.445 "base_bdevs_list": [ 00:15:53.445 { 00:15:53.445 "name": "BaseBdev1", 00:15:53.445 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:53.445 "is_configured": false, 00:15:53.445 "data_offset": 0, 00:15:53.445 "data_size": 0 00:15:53.445 }, 00:15:53.445 { 00:15:53.445 "name": "BaseBdev2", 00:15:53.445 "uuid": "0e549228-823d-407b-bdf3-6decab2c44f2", 00:15:53.445 "is_configured": true, 00:15:53.445 "data_offset": 2048, 00:15:53.445 "data_size": 63488 00:15:53.445 }, 00:15:53.445 { 00:15:53.445 "name": "BaseBdev3", 00:15:53.445 "uuid": "82beca06-afc9-413a-acae-d2668507b354", 00:15:53.445 "is_configured": true, 00:15:53.445 "data_offset": 2048, 00:15:53.445 "data_size": 63488 00:15:53.445 }, 00:15:53.445 { 00:15:53.445 "name": "BaseBdev4", 00:15:53.445 "uuid": "1be04fbd-f461-4c95-8552-8938cfcd96b0", 00:15:53.445 "is_configured": true, 00:15:53.445 "data_offset": 2048, 00:15:53.445 "data_size": 63488 00:15:53.445 } 00:15:53.445 ] 00:15:53.445 }' 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:53.445 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.703 [2024-11-20 16:03:51.844245] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:53.703 "name": "Existed_Raid", 00:15:53.703 "uuid": "3a3914f7-726f-42c3-862a-d21b1fe7fb0e", 00:15:53.703 "strip_size_kb": 64, 00:15:53.703 "state": "configuring", 00:15:53.703 "raid_level": "raid5f", 00:15:53.703 "superblock": true, 00:15:53.703 "num_base_bdevs": 4, 00:15:53.703 "num_base_bdevs_discovered": 2, 00:15:53.703 "num_base_bdevs_operational": 4, 00:15:53.703 "base_bdevs_list": [ 00:15:53.703 { 00:15:53.703 "name": "BaseBdev1", 00:15:53.703 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:53.703 "is_configured": false, 00:15:53.703 "data_offset": 0, 00:15:53.703 "data_size": 0 00:15:53.703 }, 00:15:53.703 { 00:15:53.703 "name": null, 00:15:53.703 "uuid": "0e549228-823d-407b-bdf3-6decab2c44f2", 00:15:53.703 "is_configured": false, 00:15:53.703 "data_offset": 0, 00:15:53.703 "data_size": 63488 00:15:53.703 }, 00:15:53.703 { 00:15:53.703 "name": "BaseBdev3", 00:15:53.703 "uuid": "82beca06-afc9-413a-acae-d2668507b354", 00:15:53.703 "is_configured": true, 00:15:53.703 "data_offset": 2048, 00:15:53.703 "data_size": 63488 00:15:53.703 }, 00:15:53.703 { 00:15:53.703 "name": "BaseBdev4", 00:15:53.703 "uuid": "1be04fbd-f461-4c95-8552-8938cfcd96b0", 00:15:53.703 "is_configured": true, 00:15:53.703 "data_offset": 2048, 00:15:53.703 "data_size": 63488 00:15:53.703 } 00:15:53.703 ] 00:15:53.703 }' 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:53.703 16:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.961 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.961 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.961 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.961 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:15:53.961 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.219 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:15:54.219 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:15:54.219 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.219 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.219 [2024-11-20 16:03:52.247318] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:54.219 BaseBdev1 00:15:54.219 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.219 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:15:54.219 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:15:54.219 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.220 [ 00:15:54.220 { 00:15:54.220 "name": "BaseBdev1", 00:15:54.220 "aliases": [ 00:15:54.220 "804a2c7a-7ceb-494f-a0dc-b145703d687c" 00:15:54.220 ], 00:15:54.220 "product_name": "Malloc disk", 00:15:54.220 "block_size": 512, 00:15:54.220 "num_blocks": 65536, 00:15:54.220 "uuid": "804a2c7a-7ceb-494f-a0dc-b145703d687c", 00:15:54.220 "assigned_rate_limits": { 00:15:54.220 "rw_ios_per_sec": 0, 00:15:54.220 "rw_mbytes_per_sec": 0, 00:15:54.220 "r_mbytes_per_sec": 0, 00:15:54.220 "w_mbytes_per_sec": 0 00:15:54.220 }, 00:15:54.220 "claimed": true, 00:15:54.220 "claim_type": "exclusive_write", 00:15:54.220 "zoned": false, 00:15:54.220 "supported_io_types": { 00:15:54.220 "read": true, 00:15:54.220 "write": true, 00:15:54.220 "unmap": true, 00:15:54.220 "flush": true, 00:15:54.220 "reset": true, 00:15:54.220 "nvme_admin": false, 00:15:54.220 "nvme_io": false, 00:15:54.220 "nvme_io_md": false, 00:15:54.220 "write_zeroes": true, 00:15:54.220 "zcopy": true, 00:15:54.220 "get_zone_info": false, 00:15:54.220 "zone_management": false, 00:15:54.220 "zone_append": false, 00:15:54.220 "compare": false, 00:15:54.220 "compare_and_write": false, 00:15:54.220 "abort": true, 00:15:54.220 "seek_hole": false, 00:15:54.220 "seek_data": false, 00:15:54.220 "copy": true, 00:15:54.220 "nvme_iov_md": false 00:15:54.220 }, 00:15:54.220 "memory_domains": [ 00:15:54.220 { 00:15:54.220 "dma_device_id": "system", 00:15:54.220 "dma_device_type": 1 00:15:54.220 }, 00:15:54.220 { 00:15:54.220 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:54.220 "dma_device_type": 2 00:15:54.220 } 00:15:54.220 ], 00:15:54.220 "driver_specific": {} 00:15:54.220 } 00:15:54.220 ] 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:54.220 "name": "Existed_Raid", 00:15:54.220 "uuid": "3a3914f7-726f-42c3-862a-d21b1fe7fb0e", 00:15:54.220 "strip_size_kb": 64, 00:15:54.220 "state": "configuring", 00:15:54.220 "raid_level": "raid5f", 00:15:54.220 "superblock": true, 00:15:54.220 "num_base_bdevs": 4, 00:15:54.220 "num_base_bdevs_discovered": 3, 00:15:54.220 "num_base_bdevs_operational": 4, 00:15:54.220 "base_bdevs_list": [ 00:15:54.220 { 00:15:54.220 "name": "BaseBdev1", 00:15:54.220 "uuid": "804a2c7a-7ceb-494f-a0dc-b145703d687c", 00:15:54.220 "is_configured": true, 00:15:54.220 "data_offset": 2048, 00:15:54.220 "data_size": 63488 00:15:54.220 }, 00:15:54.220 { 00:15:54.220 "name": null, 00:15:54.220 "uuid": "0e549228-823d-407b-bdf3-6decab2c44f2", 00:15:54.220 "is_configured": false, 00:15:54.220 "data_offset": 0, 00:15:54.220 "data_size": 63488 00:15:54.220 }, 00:15:54.220 { 00:15:54.220 "name": "BaseBdev3", 00:15:54.220 "uuid": "82beca06-afc9-413a-acae-d2668507b354", 00:15:54.220 "is_configured": true, 00:15:54.220 "data_offset": 2048, 00:15:54.220 "data_size": 63488 00:15:54.220 }, 00:15:54.220 { 00:15:54.220 "name": "BaseBdev4", 00:15:54.220 "uuid": "1be04fbd-f461-4c95-8552-8938cfcd96b0", 00:15:54.220 "is_configured": true, 00:15:54.220 "data_offset": 2048, 00:15:54.220 "data_size": 63488 00:15:54.220 } 00:15:54.220 ] 00:15:54.220 }' 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:54.220 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.478 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.478 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.478 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.478 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:15:54.478 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.478 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.479 [2024-11-20 16:03:52.627505] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:54.479 "name": "Existed_Raid", 00:15:54.479 "uuid": "3a3914f7-726f-42c3-862a-d21b1fe7fb0e", 00:15:54.479 "strip_size_kb": 64, 00:15:54.479 "state": "configuring", 00:15:54.479 "raid_level": "raid5f", 00:15:54.479 "superblock": true, 00:15:54.479 "num_base_bdevs": 4, 00:15:54.479 "num_base_bdevs_discovered": 2, 00:15:54.479 "num_base_bdevs_operational": 4, 00:15:54.479 "base_bdevs_list": [ 00:15:54.479 { 00:15:54.479 "name": "BaseBdev1", 00:15:54.479 "uuid": "804a2c7a-7ceb-494f-a0dc-b145703d687c", 00:15:54.479 "is_configured": true, 00:15:54.479 "data_offset": 2048, 00:15:54.479 "data_size": 63488 00:15:54.479 }, 00:15:54.479 { 00:15:54.479 "name": null, 00:15:54.479 "uuid": "0e549228-823d-407b-bdf3-6decab2c44f2", 00:15:54.479 "is_configured": false, 00:15:54.479 "data_offset": 0, 00:15:54.479 "data_size": 63488 00:15:54.479 }, 00:15:54.479 { 00:15:54.479 "name": null, 00:15:54.479 "uuid": "82beca06-afc9-413a-acae-d2668507b354", 00:15:54.479 "is_configured": false, 00:15:54.479 "data_offset": 0, 00:15:54.479 "data_size": 63488 00:15:54.479 }, 00:15:54.479 { 00:15:54.479 "name": "BaseBdev4", 00:15:54.479 "uuid": "1be04fbd-f461-4c95-8552-8938cfcd96b0", 00:15:54.479 "is_configured": true, 00:15:54.479 "data_offset": 2048, 00:15:54.479 "data_size": 63488 00:15:54.479 } 00:15:54.479 ] 00:15:54.479 }' 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:54.479 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.736 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.736 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.736 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.736 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:15:54.736 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.994 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:15:54.994 16:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:15:54.994 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.994 16:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.994 [2024-11-20 16:03:52.999570] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:54.994 "name": "Existed_Raid", 00:15:54.994 "uuid": "3a3914f7-726f-42c3-862a-d21b1fe7fb0e", 00:15:54.994 "strip_size_kb": 64, 00:15:54.994 "state": "configuring", 00:15:54.994 "raid_level": "raid5f", 00:15:54.994 "superblock": true, 00:15:54.994 "num_base_bdevs": 4, 00:15:54.994 "num_base_bdevs_discovered": 3, 00:15:54.994 "num_base_bdevs_operational": 4, 00:15:54.994 "base_bdevs_list": [ 00:15:54.994 { 00:15:54.994 "name": "BaseBdev1", 00:15:54.994 "uuid": "804a2c7a-7ceb-494f-a0dc-b145703d687c", 00:15:54.994 "is_configured": true, 00:15:54.994 "data_offset": 2048, 00:15:54.994 "data_size": 63488 00:15:54.994 }, 00:15:54.994 { 00:15:54.994 "name": null, 00:15:54.994 "uuid": "0e549228-823d-407b-bdf3-6decab2c44f2", 00:15:54.994 "is_configured": false, 00:15:54.994 "data_offset": 0, 00:15:54.994 "data_size": 63488 00:15:54.994 }, 00:15:54.994 { 00:15:54.994 "name": "BaseBdev3", 00:15:54.994 "uuid": "82beca06-afc9-413a-acae-d2668507b354", 00:15:54.994 "is_configured": true, 00:15:54.994 "data_offset": 2048, 00:15:54.994 "data_size": 63488 00:15:54.994 }, 00:15:54.994 { 00:15:54.994 "name": "BaseBdev4", 00:15:54.994 "uuid": "1be04fbd-f461-4c95-8552-8938cfcd96b0", 00:15:54.994 "is_configured": true, 00:15:54.994 "data_offset": 2048, 00:15:54.994 "data_size": 63488 00:15:54.994 } 00:15:54.994 ] 00:15:54.994 }' 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:54.994 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.277 [2024-11-20 16:03:53.339695] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:55.277 "name": "Existed_Raid", 00:15:55.277 "uuid": "3a3914f7-726f-42c3-862a-d21b1fe7fb0e", 00:15:55.277 "strip_size_kb": 64, 00:15:55.277 "state": "configuring", 00:15:55.277 "raid_level": "raid5f", 00:15:55.277 "superblock": true, 00:15:55.277 "num_base_bdevs": 4, 00:15:55.277 "num_base_bdevs_discovered": 2, 00:15:55.277 "num_base_bdevs_operational": 4, 00:15:55.277 "base_bdevs_list": [ 00:15:55.277 { 00:15:55.277 "name": null, 00:15:55.277 "uuid": "804a2c7a-7ceb-494f-a0dc-b145703d687c", 00:15:55.277 "is_configured": false, 00:15:55.277 "data_offset": 0, 00:15:55.277 "data_size": 63488 00:15:55.277 }, 00:15:55.277 { 00:15:55.277 "name": null, 00:15:55.277 "uuid": "0e549228-823d-407b-bdf3-6decab2c44f2", 00:15:55.277 "is_configured": false, 00:15:55.277 "data_offset": 0, 00:15:55.277 "data_size": 63488 00:15:55.277 }, 00:15:55.277 { 00:15:55.277 "name": "BaseBdev3", 00:15:55.277 "uuid": "82beca06-afc9-413a-acae-d2668507b354", 00:15:55.277 "is_configured": true, 00:15:55.277 "data_offset": 2048, 00:15:55.277 "data_size": 63488 00:15:55.277 }, 00:15:55.277 { 00:15:55.277 "name": "BaseBdev4", 00:15:55.277 "uuid": "1be04fbd-f461-4c95-8552-8938cfcd96b0", 00:15:55.277 "is_configured": true, 00:15:55.277 "data_offset": 2048, 00:15:55.277 "data_size": 63488 00:15:55.277 } 00:15:55.277 ] 00:15:55.277 }' 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:55.277 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.534 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:15:55.534 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.534 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.534 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.534 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.534 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:15:55.534 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:15:55.534 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.534 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.534 [2024-11-20 16:03:53.783345] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:55.791 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.791 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:55.791 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:55.791 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:55.791 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:55.791 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:55.791 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:55.791 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:55.792 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:55.792 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:55.792 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:55.792 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.792 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.792 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.792 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:55.792 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.792 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:55.792 "name": "Existed_Raid", 00:15:55.792 "uuid": "3a3914f7-726f-42c3-862a-d21b1fe7fb0e", 00:15:55.792 "strip_size_kb": 64, 00:15:55.792 "state": "configuring", 00:15:55.792 "raid_level": "raid5f", 00:15:55.792 "superblock": true, 00:15:55.792 "num_base_bdevs": 4, 00:15:55.792 "num_base_bdevs_discovered": 3, 00:15:55.792 "num_base_bdevs_operational": 4, 00:15:55.792 "base_bdevs_list": [ 00:15:55.792 { 00:15:55.792 "name": null, 00:15:55.792 "uuid": "804a2c7a-7ceb-494f-a0dc-b145703d687c", 00:15:55.792 "is_configured": false, 00:15:55.792 "data_offset": 0, 00:15:55.792 "data_size": 63488 00:15:55.792 }, 00:15:55.792 { 00:15:55.792 "name": "BaseBdev2", 00:15:55.792 "uuid": "0e549228-823d-407b-bdf3-6decab2c44f2", 00:15:55.792 "is_configured": true, 00:15:55.792 "data_offset": 2048, 00:15:55.792 "data_size": 63488 00:15:55.792 }, 00:15:55.792 { 00:15:55.792 "name": "BaseBdev3", 00:15:55.792 "uuid": "82beca06-afc9-413a-acae-d2668507b354", 00:15:55.792 "is_configured": true, 00:15:55.792 "data_offset": 2048, 00:15:55.792 "data_size": 63488 00:15:55.792 }, 00:15:55.792 { 00:15:55.792 "name": "BaseBdev4", 00:15:55.792 "uuid": "1be04fbd-f461-4c95-8552-8938cfcd96b0", 00:15:55.792 "is_configured": true, 00:15:55.792 "data_offset": 2048, 00:15:55.792 "data_size": 63488 00:15:55.792 } 00:15:55.792 ] 00:15:55.792 }' 00:15:55.792 16:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:55.792 16:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 804a2c7a-7ceb-494f-a0dc-b145703d687c 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.052 [2024-11-20 16:03:54.206148] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:15:56.052 NewBaseBdev 00:15:56.052 [2024-11-20 16:03:54.206463] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:15:56.052 [2024-11-20 16:03:54.206480] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:56.052 [2024-11-20 16:03:54.206748] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.052 [2024-11-20 16:03:54.211427] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:15:56.052 [2024-11-20 16:03:54.211448] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:15:56.052 [2024-11-20 16:03:54.211651] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.052 [ 00:15:56.052 { 00:15:56.052 "name": "NewBaseBdev", 00:15:56.052 "aliases": [ 00:15:56.052 "804a2c7a-7ceb-494f-a0dc-b145703d687c" 00:15:56.052 ], 00:15:56.052 "product_name": "Malloc disk", 00:15:56.052 "block_size": 512, 00:15:56.052 "num_blocks": 65536, 00:15:56.052 "uuid": "804a2c7a-7ceb-494f-a0dc-b145703d687c", 00:15:56.052 "assigned_rate_limits": { 00:15:56.052 "rw_ios_per_sec": 0, 00:15:56.052 "rw_mbytes_per_sec": 0, 00:15:56.052 "r_mbytes_per_sec": 0, 00:15:56.052 "w_mbytes_per_sec": 0 00:15:56.052 }, 00:15:56.052 "claimed": true, 00:15:56.052 "claim_type": "exclusive_write", 00:15:56.052 "zoned": false, 00:15:56.052 "supported_io_types": { 00:15:56.052 "read": true, 00:15:56.052 "write": true, 00:15:56.052 "unmap": true, 00:15:56.052 "flush": true, 00:15:56.052 "reset": true, 00:15:56.052 "nvme_admin": false, 00:15:56.052 "nvme_io": false, 00:15:56.052 "nvme_io_md": false, 00:15:56.052 "write_zeroes": true, 00:15:56.052 "zcopy": true, 00:15:56.052 "get_zone_info": false, 00:15:56.052 "zone_management": false, 00:15:56.052 "zone_append": false, 00:15:56.052 "compare": false, 00:15:56.052 "compare_and_write": false, 00:15:56.052 "abort": true, 00:15:56.052 "seek_hole": false, 00:15:56.052 "seek_data": false, 00:15:56.052 "copy": true, 00:15:56.052 "nvme_iov_md": false 00:15:56.052 }, 00:15:56.052 "memory_domains": [ 00:15:56.052 { 00:15:56.052 "dma_device_id": "system", 00:15:56.052 "dma_device_type": 1 00:15:56.052 }, 00:15:56.052 { 00:15:56.052 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:56.052 "dma_device_type": 2 00:15:56.052 } 00:15:56.052 ], 00:15:56.052 "driver_specific": {} 00:15:56.052 } 00:15:56.052 ] 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:56.052 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:56.053 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:56.053 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:56.053 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:56.053 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:56.053 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:56.053 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:56.053 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:56.053 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:56.053 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.053 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:56.053 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.053 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.053 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:56.053 "name": "Existed_Raid", 00:15:56.053 "uuid": "3a3914f7-726f-42c3-862a-d21b1fe7fb0e", 00:15:56.053 "strip_size_kb": 64, 00:15:56.053 "state": "online", 00:15:56.053 "raid_level": "raid5f", 00:15:56.053 "superblock": true, 00:15:56.053 "num_base_bdevs": 4, 00:15:56.053 "num_base_bdevs_discovered": 4, 00:15:56.053 "num_base_bdevs_operational": 4, 00:15:56.053 "base_bdevs_list": [ 00:15:56.053 { 00:15:56.053 "name": "NewBaseBdev", 00:15:56.053 "uuid": "804a2c7a-7ceb-494f-a0dc-b145703d687c", 00:15:56.053 "is_configured": true, 00:15:56.053 "data_offset": 2048, 00:15:56.053 "data_size": 63488 00:15:56.053 }, 00:15:56.053 { 00:15:56.053 "name": "BaseBdev2", 00:15:56.053 "uuid": "0e549228-823d-407b-bdf3-6decab2c44f2", 00:15:56.053 "is_configured": true, 00:15:56.053 "data_offset": 2048, 00:15:56.053 "data_size": 63488 00:15:56.053 }, 00:15:56.053 { 00:15:56.053 "name": "BaseBdev3", 00:15:56.053 "uuid": "82beca06-afc9-413a-acae-d2668507b354", 00:15:56.053 "is_configured": true, 00:15:56.053 "data_offset": 2048, 00:15:56.053 "data_size": 63488 00:15:56.053 }, 00:15:56.053 { 00:15:56.053 "name": "BaseBdev4", 00:15:56.053 "uuid": "1be04fbd-f461-4c95-8552-8938cfcd96b0", 00:15:56.053 "is_configured": true, 00:15:56.053 "data_offset": 2048, 00:15:56.053 "data_size": 63488 00:15:56.053 } 00:15:56.053 ] 00:15:56.053 }' 00:15:56.053 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:56.053 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.314 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:15:56.314 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:56.314 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:56.314 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:56.314 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:15:56.314 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:56.314 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:56.314 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.314 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.314 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:56.314 [2024-11-20 16:03:54.557204] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:56.575 "name": "Existed_Raid", 00:15:56.575 "aliases": [ 00:15:56.575 "3a3914f7-726f-42c3-862a-d21b1fe7fb0e" 00:15:56.575 ], 00:15:56.575 "product_name": "Raid Volume", 00:15:56.575 "block_size": 512, 00:15:56.575 "num_blocks": 190464, 00:15:56.575 "uuid": "3a3914f7-726f-42c3-862a-d21b1fe7fb0e", 00:15:56.575 "assigned_rate_limits": { 00:15:56.575 "rw_ios_per_sec": 0, 00:15:56.575 "rw_mbytes_per_sec": 0, 00:15:56.575 "r_mbytes_per_sec": 0, 00:15:56.575 "w_mbytes_per_sec": 0 00:15:56.575 }, 00:15:56.575 "claimed": false, 00:15:56.575 "zoned": false, 00:15:56.575 "supported_io_types": { 00:15:56.575 "read": true, 00:15:56.575 "write": true, 00:15:56.575 "unmap": false, 00:15:56.575 "flush": false, 00:15:56.575 "reset": true, 00:15:56.575 "nvme_admin": false, 00:15:56.575 "nvme_io": false, 00:15:56.575 "nvme_io_md": false, 00:15:56.575 "write_zeroes": true, 00:15:56.575 "zcopy": false, 00:15:56.575 "get_zone_info": false, 00:15:56.575 "zone_management": false, 00:15:56.575 "zone_append": false, 00:15:56.575 "compare": false, 00:15:56.575 "compare_and_write": false, 00:15:56.575 "abort": false, 00:15:56.575 "seek_hole": false, 00:15:56.575 "seek_data": false, 00:15:56.575 "copy": false, 00:15:56.575 "nvme_iov_md": false 00:15:56.575 }, 00:15:56.575 "driver_specific": { 00:15:56.575 "raid": { 00:15:56.575 "uuid": "3a3914f7-726f-42c3-862a-d21b1fe7fb0e", 00:15:56.575 "strip_size_kb": 64, 00:15:56.575 "state": "online", 00:15:56.575 "raid_level": "raid5f", 00:15:56.575 "superblock": true, 00:15:56.575 "num_base_bdevs": 4, 00:15:56.575 "num_base_bdevs_discovered": 4, 00:15:56.575 "num_base_bdevs_operational": 4, 00:15:56.575 "base_bdevs_list": [ 00:15:56.575 { 00:15:56.575 "name": "NewBaseBdev", 00:15:56.575 "uuid": "804a2c7a-7ceb-494f-a0dc-b145703d687c", 00:15:56.575 "is_configured": true, 00:15:56.575 "data_offset": 2048, 00:15:56.575 "data_size": 63488 00:15:56.575 }, 00:15:56.575 { 00:15:56.575 "name": "BaseBdev2", 00:15:56.575 "uuid": "0e549228-823d-407b-bdf3-6decab2c44f2", 00:15:56.575 "is_configured": true, 00:15:56.575 "data_offset": 2048, 00:15:56.575 "data_size": 63488 00:15:56.575 }, 00:15:56.575 { 00:15:56.575 "name": "BaseBdev3", 00:15:56.575 "uuid": "82beca06-afc9-413a-acae-d2668507b354", 00:15:56.575 "is_configured": true, 00:15:56.575 "data_offset": 2048, 00:15:56.575 "data_size": 63488 00:15:56.575 }, 00:15:56.575 { 00:15:56.575 "name": "BaseBdev4", 00:15:56.575 "uuid": "1be04fbd-f461-4c95-8552-8938cfcd96b0", 00:15:56.575 "is_configured": true, 00:15:56.575 "data_offset": 2048, 00:15:56.575 "data_size": 63488 00:15:56.575 } 00:15:56.575 ] 00:15:56.575 } 00:15:56.575 } 00:15:56.575 }' 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:15:56.575 BaseBdev2 00:15:56.575 BaseBdev3 00:15:56.575 BaseBdev4' 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.575 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.575 [2024-11-20 16:03:54.785037] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:56.576 [2024-11-20 16:03:54.785157] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:56.576 [2024-11-20 16:03:54.785273] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:56.576 [2024-11-20 16:03:54.785626] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:56.576 [2024-11-20 16:03:54.785728] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:15:56.576 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.576 16:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 81123 00:15:56.576 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 81123 ']' 00:15:56.576 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 81123 00:15:56.576 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:15:56.576 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:56.576 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81123 00:15:56.576 killing process with pid 81123 00:15:56.576 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:56.576 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:56.576 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81123' 00:15:56.576 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 81123 00:15:56.576 [2024-11-20 16:03:54.817429] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:56.576 16:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 81123 00:15:56.838 [2024-11-20 16:03:55.065096] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:57.783 ************************************ 00:15:57.783 END TEST raid5f_state_function_test_sb 00:15:57.783 ************************************ 00:15:57.783 16:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:15:57.783 00:15:57.783 real 0m8.655s 00:15:57.783 user 0m13.788s 00:15:57.783 sys 0m1.368s 00:15:57.783 16:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:57.783 16:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:57.783 16:03:55 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 4 00:15:57.783 16:03:55 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:15:57.783 16:03:55 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:57.783 16:03:55 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:57.783 ************************************ 00:15:57.783 START TEST raid5f_superblock_test 00:15:57.783 ************************************ 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid5f 4 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=81760 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 81760 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 81760 ']' 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:57.783 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:57.783 16:03:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:15:57.783 [2024-11-20 16:03:55.955377] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:15:57.783 [2024-11-20 16:03:55.955537] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81760 ] 00:15:58.042 [2024-11-20 16:03:56.114512] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:58.043 [2024-11-20 16:03:56.218625] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:58.302 [2024-11-20 16:03:56.355910] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:58.302 [2024-11-20 16:03:56.355956] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:58.561 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:58.561 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:15:58.561 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:15:58.561 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:58.561 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:15:58.561 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:15:58.561 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:15:58.561 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:58.561 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:58.561 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:58.561 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:15:58.561 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.562 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.838 malloc1 00:15:58.838 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.838 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:58.838 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.838 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.838 [2024-11-20 16:03:56.831112] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:58.838 [2024-11-20 16:03:56.831172] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:58.838 [2024-11-20 16:03:56.831192] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:58.838 [2024-11-20 16:03:56.831201] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:58.839 [2024-11-20 16:03:56.833372] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:58.839 [2024-11-20 16:03:56.833408] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:58.839 pt1 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.839 malloc2 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.839 [2024-11-20 16:03:56.871423] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:58.839 [2024-11-20 16:03:56.871475] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:58.839 [2024-11-20 16:03:56.871499] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:58.839 [2024-11-20 16:03:56.871508] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:58.839 [2024-11-20 16:03:56.873644] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:58.839 [2024-11-20 16:03:56.873690] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:58.839 pt2 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.839 malloc3 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.839 [2024-11-20 16:03:56.919039] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:15:58.839 [2024-11-20 16:03:56.919095] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:58.839 [2024-11-20 16:03:56.919119] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:15:58.839 [2024-11-20 16:03:56.919128] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:58.839 [2024-11-20 16:03:56.921251] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:58.839 [2024-11-20 16:03:56.921286] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:15:58.839 pt3 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.839 malloc4 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.839 [2024-11-20 16:03:56.958946] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:15:58.839 [2024-11-20 16:03:56.958996] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:58.839 [2024-11-20 16:03:56.959012] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:15:58.839 [2024-11-20 16:03:56.959020] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:58.839 [2024-11-20 16:03:56.961188] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:58.839 [2024-11-20 16:03:56.961233] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:15:58.839 pt4 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.839 [2024-11-20 16:03:56.966978] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:58.839 [2024-11-20 16:03:56.968801] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:58.839 [2024-11-20 16:03:56.968882] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:58.839 [2024-11-20 16:03:56.968928] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:15:58.839 [2024-11-20 16:03:56.969108] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:15:58.839 [2024-11-20 16:03:56.969122] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:58.839 [2024-11-20 16:03:56.969361] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:15:58.839 [2024-11-20 16:03:56.974253] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:15:58.839 [2024-11-20 16:03:56.974276] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:15:58.839 [2024-11-20 16:03:56.974444] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:58.839 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:58.840 16:03:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:58.840 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.840 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.840 16:03:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.840 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:58.840 "name": "raid_bdev1", 00:15:58.840 "uuid": "267db590-12e1-49df-999b-03ce6ed2954c", 00:15:58.840 "strip_size_kb": 64, 00:15:58.840 "state": "online", 00:15:58.840 "raid_level": "raid5f", 00:15:58.840 "superblock": true, 00:15:58.840 "num_base_bdevs": 4, 00:15:58.840 "num_base_bdevs_discovered": 4, 00:15:58.840 "num_base_bdevs_operational": 4, 00:15:58.840 "base_bdevs_list": [ 00:15:58.840 { 00:15:58.840 "name": "pt1", 00:15:58.840 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:58.840 "is_configured": true, 00:15:58.840 "data_offset": 2048, 00:15:58.840 "data_size": 63488 00:15:58.840 }, 00:15:58.840 { 00:15:58.840 "name": "pt2", 00:15:58.840 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:58.840 "is_configured": true, 00:15:58.840 "data_offset": 2048, 00:15:58.840 "data_size": 63488 00:15:58.840 }, 00:15:58.840 { 00:15:58.840 "name": "pt3", 00:15:58.840 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:58.840 "is_configured": true, 00:15:58.840 "data_offset": 2048, 00:15:58.840 "data_size": 63488 00:15:58.840 }, 00:15:58.840 { 00:15:58.840 "name": "pt4", 00:15:58.840 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:58.840 "is_configured": true, 00:15:58.840 "data_offset": 2048, 00:15:58.840 "data_size": 63488 00:15:58.840 } 00:15:58.840 ] 00:15:58.840 }' 00:15:58.840 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:58.840 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.099 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:15:59.099 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:59.099 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:59.099 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:59.099 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:15:59.099 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:59.099 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:59.099 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:59.099 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.099 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.099 [2024-11-20 16:03:57.307948] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:59.099 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.099 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:59.099 "name": "raid_bdev1", 00:15:59.099 "aliases": [ 00:15:59.099 "267db590-12e1-49df-999b-03ce6ed2954c" 00:15:59.099 ], 00:15:59.099 "product_name": "Raid Volume", 00:15:59.099 "block_size": 512, 00:15:59.099 "num_blocks": 190464, 00:15:59.099 "uuid": "267db590-12e1-49df-999b-03ce6ed2954c", 00:15:59.099 "assigned_rate_limits": { 00:15:59.099 "rw_ios_per_sec": 0, 00:15:59.099 "rw_mbytes_per_sec": 0, 00:15:59.099 "r_mbytes_per_sec": 0, 00:15:59.099 "w_mbytes_per_sec": 0 00:15:59.099 }, 00:15:59.099 "claimed": false, 00:15:59.099 "zoned": false, 00:15:59.099 "supported_io_types": { 00:15:59.099 "read": true, 00:15:59.099 "write": true, 00:15:59.099 "unmap": false, 00:15:59.099 "flush": false, 00:15:59.099 "reset": true, 00:15:59.099 "nvme_admin": false, 00:15:59.099 "nvme_io": false, 00:15:59.099 "nvme_io_md": false, 00:15:59.099 "write_zeroes": true, 00:15:59.099 "zcopy": false, 00:15:59.099 "get_zone_info": false, 00:15:59.099 "zone_management": false, 00:15:59.099 "zone_append": false, 00:15:59.099 "compare": false, 00:15:59.099 "compare_and_write": false, 00:15:59.099 "abort": false, 00:15:59.099 "seek_hole": false, 00:15:59.099 "seek_data": false, 00:15:59.099 "copy": false, 00:15:59.099 "nvme_iov_md": false 00:15:59.099 }, 00:15:59.099 "driver_specific": { 00:15:59.099 "raid": { 00:15:59.099 "uuid": "267db590-12e1-49df-999b-03ce6ed2954c", 00:15:59.099 "strip_size_kb": 64, 00:15:59.099 "state": "online", 00:15:59.099 "raid_level": "raid5f", 00:15:59.099 "superblock": true, 00:15:59.099 "num_base_bdevs": 4, 00:15:59.099 "num_base_bdevs_discovered": 4, 00:15:59.099 "num_base_bdevs_operational": 4, 00:15:59.099 "base_bdevs_list": [ 00:15:59.099 { 00:15:59.099 "name": "pt1", 00:15:59.099 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:59.099 "is_configured": true, 00:15:59.099 "data_offset": 2048, 00:15:59.099 "data_size": 63488 00:15:59.099 }, 00:15:59.099 { 00:15:59.099 "name": "pt2", 00:15:59.099 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:59.099 "is_configured": true, 00:15:59.099 "data_offset": 2048, 00:15:59.099 "data_size": 63488 00:15:59.099 }, 00:15:59.099 { 00:15:59.099 "name": "pt3", 00:15:59.099 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:59.099 "is_configured": true, 00:15:59.099 "data_offset": 2048, 00:15:59.099 "data_size": 63488 00:15:59.099 }, 00:15:59.099 { 00:15:59.099 "name": "pt4", 00:15:59.099 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:59.099 "is_configured": true, 00:15:59.099 "data_offset": 2048, 00:15:59.099 "data_size": 63488 00:15:59.099 } 00:15:59.099 ] 00:15:59.099 } 00:15:59.099 } 00:15:59.099 }' 00:15:59.099 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:59.359 pt2 00:15:59.359 pt3 00:15:59.359 pt4' 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.359 [2024-11-20 16:03:57.539987] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=267db590-12e1-49df-999b-03ce6ed2954c 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 267db590-12e1-49df-999b-03ce6ed2954c ']' 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.359 [2024-11-20 16:03:57.567807] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:59.359 [2024-11-20 16:03:57.567832] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:59.359 [2024-11-20 16:03:57.567902] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:59.359 [2024-11-20 16:03:57.567986] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:59.359 [2024-11-20 16:03:57.568000] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.359 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.618 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.619 [2024-11-20 16:03:57.679874] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:15:59.619 [2024-11-20 16:03:57.681753] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:15:59.619 [2024-11-20 16:03:57.681805] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:15:59.619 [2024-11-20 16:03:57.681839] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:15:59.619 [2024-11-20 16:03:57.681885] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:15:59.619 [2024-11-20 16:03:57.681933] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:15:59.619 [2024-11-20 16:03:57.681952] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:15:59.619 [2024-11-20 16:03:57.681970] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:15:59.619 [2024-11-20 16:03:57.681984] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:59.619 [2024-11-20 16:03:57.681995] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:15:59.619 request: 00:15:59.619 { 00:15:59.619 "name": "raid_bdev1", 00:15:59.619 "raid_level": "raid5f", 00:15:59.619 "base_bdevs": [ 00:15:59.619 "malloc1", 00:15:59.619 "malloc2", 00:15:59.619 "malloc3", 00:15:59.619 "malloc4" 00:15:59.619 ], 00:15:59.619 "strip_size_kb": 64, 00:15:59.619 "superblock": false, 00:15:59.619 "method": "bdev_raid_create", 00:15:59.619 "req_id": 1 00:15:59.619 } 00:15:59.619 Got JSON-RPC error response 00:15:59.619 response: 00:15:59.619 { 00:15:59.619 "code": -17, 00:15:59.619 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:15:59.619 } 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.619 [2024-11-20 16:03:57.727835] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:59.619 [2024-11-20 16:03:57.727892] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:59.619 [2024-11-20 16:03:57.727910] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:15:59.619 [2024-11-20 16:03:57.727921] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:59.619 [2024-11-20 16:03:57.730079] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:59.619 [2024-11-20 16:03:57.730116] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:59.619 [2024-11-20 16:03:57.730189] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:59.619 [2024-11-20 16:03:57.730237] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:59.619 pt1 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:59.619 "name": "raid_bdev1", 00:15:59.619 "uuid": "267db590-12e1-49df-999b-03ce6ed2954c", 00:15:59.619 "strip_size_kb": 64, 00:15:59.619 "state": "configuring", 00:15:59.619 "raid_level": "raid5f", 00:15:59.619 "superblock": true, 00:15:59.619 "num_base_bdevs": 4, 00:15:59.619 "num_base_bdevs_discovered": 1, 00:15:59.619 "num_base_bdevs_operational": 4, 00:15:59.619 "base_bdevs_list": [ 00:15:59.619 { 00:15:59.619 "name": "pt1", 00:15:59.619 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:59.619 "is_configured": true, 00:15:59.619 "data_offset": 2048, 00:15:59.619 "data_size": 63488 00:15:59.619 }, 00:15:59.619 { 00:15:59.619 "name": null, 00:15:59.619 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:59.619 "is_configured": false, 00:15:59.619 "data_offset": 2048, 00:15:59.619 "data_size": 63488 00:15:59.619 }, 00:15:59.619 { 00:15:59.619 "name": null, 00:15:59.619 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:59.619 "is_configured": false, 00:15:59.619 "data_offset": 2048, 00:15:59.619 "data_size": 63488 00:15:59.619 }, 00:15:59.619 { 00:15:59.619 "name": null, 00:15:59.619 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:59.619 "is_configured": false, 00:15:59.619 "data_offset": 2048, 00:15:59.619 "data_size": 63488 00:15:59.619 } 00:15:59.619 ] 00:15:59.619 }' 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:59.619 16:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.878 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:15:59.878 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:59.878 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.878 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.878 [2024-11-20 16:03:58.055936] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:59.878 [2024-11-20 16:03:58.056006] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:59.878 [2024-11-20 16:03:58.056025] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:15:59.878 [2024-11-20 16:03:58.056035] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:59.878 [2024-11-20 16:03:58.056451] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:59.878 [2024-11-20 16:03:58.056467] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:59.878 [2024-11-20 16:03:58.056537] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:59.878 [2024-11-20 16:03:58.056559] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:59.878 pt2 00:15:59.878 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.878 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:15:59.878 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.878 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.878 [2024-11-20 16:03:58.063941] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:15:59.878 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.878 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:15:59.878 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:59.878 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:59.878 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:59.878 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:59.879 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:59.879 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:59.879 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:59.879 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:59.879 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:59.879 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:59.879 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:59.879 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.879 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.879 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.879 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:59.879 "name": "raid_bdev1", 00:15:59.879 "uuid": "267db590-12e1-49df-999b-03ce6ed2954c", 00:15:59.879 "strip_size_kb": 64, 00:15:59.879 "state": "configuring", 00:15:59.879 "raid_level": "raid5f", 00:15:59.879 "superblock": true, 00:15:59.879 "num_base_bdevs": 4, 00:15:59.879 "num_base_bdevs_discovered": 1, 00:15:59.879 "num_base_bdevs_operational": 4, 00:15:59.879 "base_bdevs_list": [ 00:15:59.879 { 00:15:59.879 "name": "pt1", 00:15:59.879 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:59.879 "is_configured": true, 00:15:59.879 "data_offset": 2048, 00:15:59.879 "data_size": 63488 00:15:59.879 }, 00:15:59.879 { 00:15:59.879 "name": null, 00:15:59.879 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:59.879 "is_configured": false, 00:15:59.879 "data_offset": 0, 00:15:59.879 "data_size": 63488 00:15:59.879 }, 00:15:59.879 { 00:15:59.879 "name": null, 00:15:59.879 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:59.879 "is_configured": false, 00:15:59.879 "data_offset": 2048, 00:15:59.879 "data_size": 63488 00:15:59.879 }, 00:15:59.879 { 00:15:59.879 "name": null, 00:15:59.879 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:59.879 "is_configured": false, 00:15:59.879 "data_offset": 2048, 00:15:59.879 "data_size": 63488 00:15:59.879 } 00:15:59.879 ] 00:15:59.879 }' 00:15:59.879 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:59.879 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.447 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:16:00.447 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:00.447 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:00.447 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.447 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.447 [2024-11-20 16:03:58.420024] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:00.447 [2024-11-20 16:03:58.420083] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:00.447 [2024-11-20 16:03:58.420102] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:16:00.447 [2024-11-20 16:03:58.420111] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:00.447 [2024-11-20 16:03:58.420525] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:00.447 [2024-11-20 16:03:58.420539] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:00.447 [2024-11-20 16:03:58.420609] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:00.447 [2024-11-20 16:03:58.420628] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:00.447 pt2 00:16:00.447 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.447 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:16:00.447 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:00.447 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:16:00.447 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.447 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.447 [2024-11-20 16:03:58.428000] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:16:00.447 [2024-11-20 16:03:58.428044] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:00.447 [2024-11-20 16:03:58.428063] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:16:00.447 [2024-11-20 16:03:58.428071] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:00.447 [2024-11-20 16:03:58.428447] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:00.447 [2024-11-20 16:03:58.428467] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:16:00.447 [2024-11-20 16:03:58.428525] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:16:00.447 [2024-11-20 16:03:58.428545] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:16:00.447 pt3 00:16:00.447 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.447 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.448 [2024-11-20 16:03:58.435983] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:16:00.448 [2024-11-20 16:03:58.436022] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:00.448 [2024-11-20 16:03:58.436037] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:16:00.448 [2024-11-20 16:03:58.436045] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:00.448 [2024-11-20 16:03:58.436397] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:00.448 [2024-11-20 16:03:58.436418] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:16:00.448 [2024-11-20 16:03:58.436475] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:16:00.448 [2024-11-20 16:03:58.436492] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:16:00.448 [2024-11-20 16:03:58.436621] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:16:00.448 [2024-11-20 16:03:58.436630] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:16:00.448 [2024-11-20 16:03:58.436876] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:16:00.448 [2024-11-20 16:03:58.441436] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:16:00.448 [2024-11-20 16:03:58.441459] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:16:00.448 [2024-11-20 16:03:58.441619] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:00.448 pt4 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:00.448 "name": "raid_bdev1", 00:16:00.448 "uuid": "267db590-12e1-49df-999b-03ce6ed2954c", 00:16:00.448 "strip_size_kb": 64, 00:16:00.448 "state": "online", 00:16:00.448 "raid_level": "raid5f", 00:16:00.448 "superblock": true, 00:16:00.448 "num_base_bdevs": 4, 00:16:00.448 "num_base_bdevs_discovered": 4, 00:16:00.448 "num_base_bdevs_operational": 4, 00:16:00.448 "base_bdevs_list": [ 00:16:00.448 { 00:16:00.448 "name": "pt1", 00:16:00.448 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:00.448 "is_configured": true, 00:16:00.448 "data_offset": 2048, 00:16:00.448 "data_size": 63488 00:16:00.448 }, 00:16:00.448 { 00:16:00.448 "name": "pt2", 00:16:00.448 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:00.448 "is_configured": true, 00:16:00.448 "data_offset": 2048, 00:16:00.448 "data_size": 63488 00:16:00.448 }, 00:16:00.448 { 00:16:00.448 "name": "pt3", 00:16:00.448 "uuid": "00000000-0000-0000-0000-000000000003", 00:16:00.448 "is_configured": true, 00:16:00.448 "data_offset": 2048, 00:16:00.448 "data_size": 63488 00:16:00.448 }, 00:16:00.448 { 00:16:00.448 "name": "pt4", 00:16:00.448 "uuid": "00000000-0000-0000-0000-000000000004", 00:16:00.448 "is_configured": true, 00:16:00.448 "data_offset": 2048, 00:16:00.448 "data_size": 63488 00:16:00.448 } 00:16:00.448 ] 00:16:00.448 }' 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:00.448 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.705 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:16:00.705 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:00.705 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:00.705 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:00.705 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:16:00.705 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:00.705 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:00.705 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:00.705 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.705 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.705 [2024-11-20 16:03:58.763154] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:00.706 "name": "raid_bdev1", 00:16:00.706 "aliases": [ 00:16:00.706 "267db590-12e1-49df-999b-03ce6ed2954c" 00:16:00.706 ], 00:16:00.706 "product_name": "Raid Volume", 00:16:00.706 "block_size": 512, 00:16:00.706 "num_blocks": 190464, 00:16:00.706 "uuid": "267db590-12e1-49df-999b-03ce6ed2954c", 00:16:00.706 "assigned_rate_limits": { 00:16:00.706 "rw_ios_per_sec": 0, 00:16:00.706 "rw_mbytes_per_sec": 0, 00:16:00.706 "r_mbytes_per_sec": 0, 00:16:00.706 "w_mbytes_per_sec": 0 00:16:00.706 }, 00:16:00.706 "claimed": false, 00:16:00.706 "zoned": false, 00:16:00.706 "supported_io_types": { 00:16:00.706 "read": true, 00:16:00.706 "write": true, 00:16:00.706 "unmap": false, 00:16:00.706 "flush": false, 00:16:00.706 "reset": true, 00:16:00.706 "nvme_admin": false, 00:16:00.706 "nvme_io": false, 00:16:00.706 "nvme_io_md": false, 00:16:00.706 "write_zeroes": true, 00:16:00.706 "zcopy": false, 00:16:00.706 "get_zone_info": false, 00:16:00.706 "zone_management": false, 00:16:00.706 "zone_append": false, 00:16:00.706 "compare": false, 00:16:00.706 "compare_and_write": false, 00:16:00.706 "abort": false, 00:16:00.706 "seek_hole": false, 00:16:00.706 "seek_data": false, 00:16:00.706 "copy": false, 00:16:00.706 "nvme_iov_md": false 00:16:00.706 }, 00:16:00.706 "driver_specific": { 00:16:00.706 "raid": { 00:16:00.706 "uuid": "267db590-12e1-49df-999b-03ce6ed2954c", 00:16:00.706 "strip_size_kb": 64, 00:16:00.706 "state": "online", 00:16:00.706 "raid_level": "raid5f", 00:16:00.706 "superblock": true, 00:16:00.706 "num_base_bdevs": 4, 00:16:00.706 "num_base_bdevs_discovered": 4, 00:16:00.706 "num_base_bdevs_operational": 4, 00:16:00.706 "base_bdevs_list": [ 00:16:00.706 { 00:16:00.706 "name": "pt1", 00:16:00.706 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:00.706 "is_configured": true, 00:16:00.706 "data_offset": 2048, 00:16:00.706 "data_size": 63488 00:16:00.706 }, 00:16:00.706 { 00:16:00.706 "name": "pt2", 00:16:00.706 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:00.706 "is_configured": true, 00:16:00.706 "data_offset": 2048, 00:16:00.706 "data_size": 63488 00:16:00.706 }, 00:16:00.706 { 00:16:00.706 "name": "pt3", 00:16:00.706 "uuid": "00000000-0000-0000-0000-000000000003", 00:16:00.706 "is_configured": true, 00:16:00.706 "data_offset": 2048, 00:16:00.706 "data_size": 63488 00:16:00.706 }, 00:16:00.706 { 00:16:00.706 "name": "pt4", 00:16:00.706 "uuid": "00000000-0000-0000-0000-000000000004", 00:16:00.706 "is_configured": true, 00:16:00.706 "data_offset": 2048, 00:16:00.706 "data_size": 63488 00:16:00.706 } 00:16:00.706 ] 00:16:00.706 } 00:16:00.706 } 00:16:00.706 }' 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:00.706 pt2 00:16:00.706 pt3 00:16:00.706 pt4' 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.706 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.965 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:16:00.965 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:16:00.965 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:00.965 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:00.965 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:16:00.965 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.965 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.965 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.965 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:16:00.965 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:16:00.965 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:00.965 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.965 16:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.965 16:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:16:00.965 [2024-11-20 16:03:58.991183] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 267db590-12e1-49df-999b-03ce6ed2954c '!=' 267db590-12e1-49df-999b-03ce6ed2954c ']' 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.965 [2024-11-20 16:03:59.027049] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:00.965 "name": "raid_bdev1", 00:16:00.965 "uuid": "267db590-12e1-49df-999b-03ce6ed2954c", 00:16:00.965 "strip_size_kb": 64, 00:16:00.965 "state": "online", 00:16:00.965 "raid_level": "raid5f", 00:16:00.965 "superblock": true, 00:16:00.965 "num_base_bdevs": 4, 00:16:00.965 "num_base_bdevs_discovered": 3, 00:16:00.965 "num_base_bdevs_operational": 3, 00:16:00.965 "base_bdevs_list": [ 00:16:00.965 { 00:16:00.965 "name": null, 00:16:00.965 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:00.965 "is_configured": false, 00:16:00.965 "data_offset": 0, 00:16:00.965 "data_size": 63488 00:16:00.965 }, 00:16:00.965 { 00:16:00.965 "name": "pt2", 00:16:00.965 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:00.965 "is_configured": true, 00:16:00.965 "data_offset": 2048, 00:16:00.965 "data_size": 63488 00:16:00.965 }, 00:16:00.965 { 00:16:00.965 "name": "pt3", 00:16:00.965 "uuid": "00000000-0000-0000-0000-000000000003", 00:16:00.965 "is_configured": true, 00:16:00.965 "data_offset": 2048, 00:16:00.965 "data_size": 63488 00:16:00.965 }, 00:16:00.965 { 00:16:00.965 "name": "pt4", 00:16:00.965 "uuid": "00000000-0000-0000-0000-000000000004", 00:16:00.965 "is_configured": true, 00:16:00.965 "data_offset": 2048, 00:16:00.965 "data_size": 63488 00:16:00.965 } 00:16:00.965 ] 00:16:00.965 }' 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:00.965 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.222 [2024-11-20 16:03:59.367084] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:01.222 [2024-11-20 16:03:59.367114] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:01.222 [2024-11-20 16:03:59.367178] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:01.222 [2024-11-20 16:03:59.367254] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:01.222 [2024-11-20 16:03:59.367263] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.222 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.222 [2024-11-20 16:03:59.435110] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:01.223 [2024-11-20 16:03:59.435167] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:01.223 [2024-11-20 16:03:59.435185] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:16:01.223 [2024-11-20 16:03:59.435194] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:01.223 [2024-11-20 16:03:59.437427] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:01.223 [2024-11-20 16:03:59.437575] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:01.223 [2024-11-20 16:03:59.437684] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:01.223 [2024-11-20 16:03:59.437732] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:01.223 pt2 00:16:01.223 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.223 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:16:01.223 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:01.223 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:01.223 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:01.223 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:01.223 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:01.223 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:01.223 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:01.223 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:01.223 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:01.223 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:01.223 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.223 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:01.223 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.223 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.480 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:01.480 "name": "raid_bdev1", 00:16:01.480 "uuid": "267db590-12e1-49df-999b-03ce6ed2954c", 00:16:01.480 "strip_size_kb": 64, 00:16:01.480 "state": "configuring", 00:16:01.480 "raid_level": "raid5f", 00:16:01.480 "superblock": true, 00:16:01.480 "num_base_bdevs": 4, 00:16:01.480 "num_base_bdevs_discovered": 1, 00:16:01.480 "num_base_bdevs_operational": 3, 00:16:01.480 "base_bdevs_list": [ 00:16:01.480 { 00:16:01.480 "name": null, 00:16:01.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:01.480 "is_configured": false, 00:16:01.480 "data_offset": 2048, 00:16:01.480 "data_size": 63488 00:16:01.480 }, 00:16:01.480 { 00:16:01.480 "name": "pt2", 00:16:01.480 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:01.480 "is_configured": true, 00:16:01.480 "data_offset": 2048, 00:16:01.480 "data_size": 63488 00:16:01.480 }, 00:16:01.480 { 00:16:01.480 "name": null, 00:16:01.480 "uuid": "00000000-0000-0000-0000-000000000003", 00:16:01.480 "is_configured": false, 00:16:01.480 "data_offset": 2048, 00:16:01.480 "data_size": 63488 00:16:01.480 }, 00:16:01.480 { 00:16:01.480 "name": null, 00:16:01.480 "uuid": "00000000-0000-0000-0000-000000000004", 00:16:01.480 "is_configured": false, 00:16:01.480 "data_offset": 2048, 00:16:01.480 "data_size": 63488 00:16:01.480 } 00:16:01.480 ] 00:16:01.480 }' 00:16:01.480 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:01.480 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.738 [2024-11-20 16:03:59.767185] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:16:01.738 [2024-11-20 16:03:59.767252] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:01.738 [2024-11-20 16:03:59.767270] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:16:01.738 [2024-11-20 16:03:59.767279] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:01.738 [2024-11-20 16:03:59.767658] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:01.738 [2024-11-20 16:03:59.767695] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:16:01.738 [2024-11-20 16:03:59.767765] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:16:01.738 [2024-11-20 16:03:59.767785] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:16:01.738 pt3 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:01.738 "name": "raid_bdev1", 00:16:01.738 "uuid": "267db590-12e1-49df-999b-03ce6ed2954c", 00:16:01.738 "strip_size_kb": 64, 00:16:01.738 "state": "configuring", 00:16:01.738 "raid_level": "raid5f", 00:16:01.738 "superblock": true, 00:16:01.738 "num_base_bdevs": 4, 00:16:01.738 "num_base_bdevs_discovered": 2, 00:16:01.738 "num_base_bdevs_operational": 3, 00:16:01.738 "base_bdevs_list": [ 00:16:01.738 { 00:16:01.738 "name": null, 00:16:01.738 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:01.738 "is_configured": false, 00:16:01.738 "data_offset": 2048, 00:16:01.738 "data_size": 63488 00:16:01.738 }, 00:16:01.738 { 00:16:01.738 "name": "pt2", 00:16:01.738 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:01.738 "is_configured": true, 00:16:01.738 "data_offset": 2048, 00:16:01.738 "data_size": 63488 00:16:01.738 }, 00:16:01.738 { 00:16:01.738 "name": "pt3", 00:16:01.738 "uuid": "00000000-0000-0000-0000-000000000003", 00:16:01.738 "is_configured": true, 00:16:01.738 "data_offset": 2048, 00:16:01.738 "data_size": 63488 00:16:01.738 }, 00:16:01.738 { 00:16:01.738 "name": null, 00:16:01.738 "uuid": "00000000-0000-0000-0000-000000000004", 00:16:01.738 "is_configured": false, 00:16:01.738 "data_offset": 2048, 00:16:01.738 "data_size": 63488 00:16:01.738 } 00:16:01.738 ] 00:16:01.738 }' 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:01.738 16:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.997 [2024-11-20 16:04:00.103287] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:16:01.997 [2024-11-20 16:04:00.103347] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:01.997 [2024-11-20 16:04:00.103365] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:16:01.997 [2024-11-20 16:04:00.103374] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:01.997 [2024-11-20 16:04:00.103799] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:01.997 [2024-11-20 16:04:00.103820] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:16:01.997 [2024-11-20 16:04:00.103891] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:16:01.997 [2024-11-20 16:04:00.103952] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:16:01.997 [2024-11-20 16:04:00.104082] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:16:01.997 [2024-11-20 16:04:00.104091] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:16:01.997 [2024-11-20 16:04:00.104332] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:16:01.997 [2024-11-20 16:04:00.108946] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:16:01.997 [2024-11-20 16:04:00.108969] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:16:01.997 [2024-11-20 16:04:00.109263] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:01.997 pt4 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:01.997 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.998 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.998 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:01.998 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.998 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:01.998 "name": "raid_bdev1", 00:16:01.998 "uuid": "267db590-12e1-49df-999b-03ce6ed2954c", 00:16:01.998 "strip_size_kb": 64, 00:16:01.998 "state": "online", 00:16:01.998 "raid_level": "raid5f", 00:16:01.998 "superblock": true, 00:16:01.998 "num_base_bdevs": 4, 00:16:01.998 "num_base_bdevs_discovered": 3, 00:16:01.998 "num_base_bdevs_operational": 3, 00:16:01.998 "base_bdevs_list": [ 00:16:01.998 { 00:16:01.998 "name": null, 00:16:01.998 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:01.998 "is_configured": false, 00:16:01.998 "data_offset": 2048, 00:16:01.998 "data_size": 63488 00:16:01.998 }, 00:16:01.998 { 00:16:01.998 "name": "pt2", 00:16:01.998 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:01.998 "is_configured": true, 00:16:01.998 "data_offset": 2048, 00:16:01.998 "data_size": 63488 00:16:01.998 }, 00:16:01.998 { 00:16:01.998 "name": "pt3", 00:16:01.998 "uuid": "00000000-0000-0000-0000-000000000003", 00:16:01.998 "is_configured": true, 00:16:01.998 "data_offset": 2048, 00:16:01.998 "data_size": 63488 00:16:01.998 }, 00:16:01.998 { 00:16:01.998 "name": "pt4", 00:16:01.998 "uuid": "00000000-0000-0000-0000-000000000004", 00:16:01.998 "is_configured": true, 00:16:01.998 "data_offset": 2048, 00:16:01.998 "data_size": 63488 00:16:01.998 } 00:16:01.998 ] 00:16:01.998 }' 00:16:01.998 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:01.998 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.258 [2024-11-20 16:04:00.442577] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:02.258 [2024-11-20 16:04:00.442600] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:02.258 [2024-11-20 16:04:00.442659] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:02.258 [2024-11-20 16:04:00.442743] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:02.258 [2024-11-20 16:04:00.442756] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.258 [2024-11-20 16:04:00.494598] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:02.258 [2024-11-20 16:04:00.494660] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:02.258 [2024-11-20 16:04:00.494689] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:16:02.258 [2024-11-20 16:04:00.494703] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:02.258 [2024-11-20 16:04:00.496956] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:02.258 [2024-11-20 16:04:00.496993] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:02.258 [2024-11-20 16:04:00.497074] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:02.258 [2024-11-20 16:04:00.497115] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:02.258 [2024-11-20 16:04:00.497239] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:16:02.258 [2024-11-20 16:04:00.497253] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:02.258 [2024-11-20 16:04:00.497267] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:16:02.258 [2024-11-20 16:04:00.497320] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:02.258 [2024-11-20 16:04:00.497414] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:16:02.258 pt1 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.258 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:02.516 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.516 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:02.516 "name": "raid_bdev1", 00:16:02.516 "uuid": "267db590-12e1-49df-999b-03ce6ed2954c", 00:16:02.516 "strip_size_kb": 64, 00:16:02.516 "state": "configuring", 00:16:02.516 "raid_level": "raid5f", 00:16:02.516 "superblock": true, 00:16:02.516 "num_base_bdevs": 4, 00:16:02.516 "num_base_bdevs_discovered": 2, 00:16:02.516 "num_base_bdevs_operational": 3, 00:16:02.516 "base_bdevs_list": [ 00:16:02.516 { 00:16:02.516 "name": null, 00:16:02.516 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:02.516 "is_configured": false, 00:16:02.516 "data_offset": 2048, 00:16:02.516 "data_size": 63488 00:16:02.516 }, 00:16:02.516 { 00:16:02.516 "name": "pt2", 00:16:02.516 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:02.516 "is_configured": true, 00:16:02.516 "data_offset": 2048, 00:16:02.516 "data_size": 63488 00:16:02.516 }, 00:16:02.516 { 00:16:02.516 "name": "pt3", 00:16:02.516 "uuid": "00000000-0000-0000-0000-000000000003", 00:16:02.516 "is_configured": true, 00:16:02.516 "data_offset": 2048, 00:16:02.516 "data_size": 63488 00:16:02.516 }, 00:16:02.516 { 00:16:02.516 "name": null, 00:16:02.516 "uuid": "00000000-0000-0000-0000-000000000004", 00:16:02.516 "is_configured": false, 00:16:02.516 "data_offset": 2048, 00:16:02.516 "data_size": 63488 00:16:02.516 } 00:16:02.516 ] 00:16:02.516 }' 00:16:02.516 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:02.516 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.782 [2024-11-20 16:04:00.858715] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:16:02.782 [2024-11-20 16:04:00.858771] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:02.782 [2024-11-20 16:04:00.858791] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:16:02.782 [2024-11-20 16:04:00.858800] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:02.782 [2024-11-20 16:04:00.859197] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:02.782 [2024-11-20 16:04:00.859229] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:16:02.782 [2024-11-20 16:04:00.859313] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:16:02.782 [2024-11-20 16:04:00.859335] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:16:02.782 [2024-11-20 16:04:00.859459] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:16:02.782 [2024-11-20 16:04:00.859471] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:16:02.782 [2024-11-20 16:04:00.859719] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:16:02.782 [2024-11-20 16:04:00.864402] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:16:02.782 [2024-11-20 16:04:00.864424] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:16:02.782 [2024-11-20 16:04:00.864684] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:02.782 pt4 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:02.782 "name": "raid_bdev1", 00:16:02.782 "uuid": "267db590-12e1-49df-999b-03ce6ed2954c", 00:16:02.782 "strip_size_kb": 64, 00:16:02.782 "state": "online", 00:16:02.782 "raid_level": "raid5f", 00:16:02.782 "superblock": true, 00:16:02.782 "num_base_bdevs": 4, 00:16:02.782 "num_base_bdevs_discovered": 3, 00:16:02.782 "num_base_bdevs_operational": 3, 00:16:02.782 "base_bdevs_list": [ 00:16:02.782 { 00:16:02.782 "name": null, 00:16:02.782 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:02.782 "is_configured": false, 00:16:02.782 "data_offset": 2048, 00:16:02.782 "data_size": 63488 00:16:02.782 }, 00:16:02.782 { 00:16:02.782 "name": "pt2", 00:16:02.782 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:02.782 "is_configured": true, 00:16:02.782 "data_offset": 2048, 00:16:02.782 "data_size": 63488 00:16:02.782 }, 00:16:02.782 { 00:16:02.782 "name": "pt3", 00:16:02.782 "uuid": "00000000-0000-0000-0000-000000000003", 00:16:02.782 "is_configured": true, 00:16:02.782 "data_offset": 2048, 00:16:02.782 "data_size": 63488 00:16:02.782 }, 00:16:02.782 { 00:16:02.782 "name": "pt4", 00:16:02.782 "uuid": "00000000-0000-0000-0000-000000000004", 00:16:02.782 "is_configured": true, 00:16:02.782 "data_offset": 2048, 00:16:02.782 "data_size": 63488 00:16:02.782 } 00:16:02.782 ] 00:16:02.782 }' 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:02.782 16:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:03.040 [2024-11-20 16:04:01.210215] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 267db590-12e1-49df-999b-03ce6ed2954c '!=' 267db590-12e1-49df-999b-03ce6ed2954c ']' 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 81760 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 81760 ']' 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # kill -0 81760 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # uname 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81760 00:16:03.040 killing process with pid 81760 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81760' 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@973 -- # kill 81760 00:16:03.040 [2024-11-20 16:04:01.260971] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:03.040 [2024-11-20 16:04:01.261054] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:03.040 16:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@978 -- # wait 81760 00:16:03.040 [2024-11-20 16:04:01.261126] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:03.040 [2024-11-20 16:04:01.261138] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:16:03.298 [2024-11-20 16:04:01.510927] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:04.229 16:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:16:04.229 00:16:04.229 real 0m6.347s 00:16:04.229 user 0m10.005s 00:16:04.229 sys 0m1.026s 00:16:04.229 16:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:04.230 ************************************ 00:16:04.230 END TEST raid5f_superblock_test 00:16:04.230 ************************************ 00:16:04.230 16:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:04.230 16:04:02 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:16:04.230 16:04:02 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 4 false false true 00:16:04.230 16:04:02 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:16:04.230 16:04:02 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:04.230 16:04:02 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:04.230 ************************************ 00:16:04.230 START TEST raid5f_rebuild_test 00:16:04.230 ************************************ 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 4 false false true 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:04.230 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=82224 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 82224 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 82224 ']' 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:04.230 16:04:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:16:04.230 I/O size of 3145728 is greater than zero copy threshold (65536). 00:16:04.230 Zero copy mechanism will not be used. 00:16:04.230 [2024-11-20 16:04:02.347320] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:16:04.230 [2024-11-20 16:04:02.347441] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid82224 ] 00:16:04.487 [2024-11-20 16:04:02.505259] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:04.487 [2024-11-20 16:04:02.609831] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:04.744 [2024-11-20 16:04:02.748144] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:04.744 [2024-11-20 16:04:02.748185] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:05.002 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:05.002 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:16:05.002 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:05.002 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:16:05.002 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.002 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.002 BaseBdev1_malloc 00:16:05.002 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.002 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:05.002 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.002 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.002 [2024-11-20 16:04:03.227807] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:05.002 [2024-11-20 16:04:03.227868] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:05.002 [2024-11-20 16:04:03.227890] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:05.002 [2024-11-20 16:04:03.227901] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:05.002 [2024-11-20 16:04:03.230463] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:05.002 [2024-11-20 16:04:03.230513] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:05.002 BaseBdev1 00:16:05.002 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.002 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:05.002 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:16:05.002 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.002 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.302 BaseBdev2_malloc 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.302 [2024-11-20 16:04:03.265758] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:16:05.302 [2024-11-20 16:04:03.265939] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:05.302 [2024-11-20 16:04:03.265968] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:05.302 [2024-11-20 16:04:03.265979] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:05.302 [2024-11-20 16:04:03.268120] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:05.302 [2024-11-20 16:04:03.268156] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:05.302 BaseBdev2 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.302 BaseBdev3_malloc 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.302 [2024-11-20 16:04:03.322067] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:16:05.302 [2024-11-20 16:04:03.322127] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:05.302 [2024-11-20 16:04:03.322150] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:16:05.302 [2024-11-20 16:04:03.322161] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:05.302 [2024-11-20 16:04:03.324446] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:05.302 [2024-11-20 16:04:03.324488] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:16:05.302 BaseBdev3 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.302 BaseBdev4_malloc 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.302 [2024-11-20 16:04:03.363085] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:16:05.302 [2024-11-20 16:04:03.363160] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:05.302 [2024-11-20 16:04:03.363184] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:16:05.302 [2024-11-20 16:04:03.363198] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:05.302 [2024-11-20 16:04:03.365466] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:05.302 [2024-11-20 16:04:03.365624] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:16:05.302 BaseBdev4 00:16:05.302 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.303 spare_malloc 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.303 spare_delay 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.303 [2024-11-20 16:04:03.412019] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:05.303 [2024-11-20 16:04:03.412189] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:05.303 [2024-11-20 16:04:03.412214] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:16:05.303 [2024-11-20 16:04:03.412225] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:05.303 [2024-11-20 16:04:03.414368] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:05.303 [2024-11-20 16:04:03.414403] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:05.303 spare 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.303 [2024-11-20 16:04:03.420073] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:05.303 [2024-11-20 16:04:03.421897] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:05.303 [2024-11-20 16:04:03.421952] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:16:05.303 [2024-11-20 16:04:03.422002] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:16:05.303 [2024-11-20 16:04:03.422087] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:16:05.303 [2024-11-20 16:04:03.422099] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:16:05.303 [2024-11-20 16:04:03.422396] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:16:05.303 [2024-11-20 16:04:03.427334] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:16:05.303 [2024-11-20 16:04:03.427358] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:16:05.303 [2024-11-20 16:04:03.427561] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:05.303 "name": "raid_bdev1", 00:16:05.303 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:05.303 "strip_size_kb": 64, 00:16:05.303 "state": "online", 00:16:05.303 "raid_level": "raid5f", 00:16:05.303 "superblock": false, 00:16:05.303 "num_base_bdevs": 4, 00:16:05.303 "num_base_bdevs_discovered": 4, 00:16:05.303 "num_base_bdevs_operational": 4, 00:16:05.303 "base_bdevs_list": [ 00:16:05.303 { 00:16:05.303 "name": "BaseBdev1", 00:16:05.303 "uuid": "c3180713-635b-5c41-ac1d-4afaef976317", 00:16:05.303 "is_configured": true, 00:16:05.303 "data_offset": 0, 00:16:05.303 "data_size": 65536 00:16:05.303 }, 00:16:05.303 { 00:16:05.303 "name": "BaseBdev2", 00:16:05.303 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:05.303 "is_configured": true, 00:16:05.303 "data_offset": 0, 00:16:05.303 "data_size": 65536 00:16:05.303 }, 00:16:05.303 { 00:16:05.303 "name": "BaseBdev3", 00:16:05.303 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:05.303 "is_configured": true, 00:16:05.303 "data_offset": 0, 00:16:05.303 "data_size": 65536 00:16:05.303 }, 00:16:05.303 { 00:16:05.303 "name": "BaseBdev4", 00:16:05.303 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:05.303 "is_configured": true, 00:16:05.303 "data_offset": 0, 00:16:05.303 "data_size": 65536 00:16:05.303 } 00:16:05.303 ] 00:16:05.303 }' 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:05.303 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.560 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:05.560 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:16:05.560 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.560 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.560 [2024-11-20 16:04:03.769337] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:05.560 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.560 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=196608 00:16:05.561 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:05.561 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:16:05.561 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.561 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.817 16:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.817 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:16:05.818 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:16:05.818 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:16:05.818 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:16:05.818 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:16:05.818 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:05.818 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:16:05.818 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:05.818 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:16:05.818 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:05.818 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:16:05.818 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:05.818 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:05.818 16:04:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:16:05.818 [2024-11-20 16:04:04.021218] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:16:05.818 /dev/nbd0 00:16:05.818 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:05.818 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:05.818 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:05.818 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:16:05.818 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:05.818 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:05.818 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:05.818 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:16:05.818 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:05.818 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:05.818 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:05.818 1+0 records in 00:16:05.818 1+0 records out 00:16:05.818 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000266358 s, 15.4 MB/s 00:16:06.075 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:06.075 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:16:06.075 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:06.075 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:06.075 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:16:06.075 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:06.075 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:06.075 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:16:06.075 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:16:06.075 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 192 00:16:06.075 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=512 oflag=direct 00:16:06.641 512+0 records in 00:16:06.641 512+0 records out 00:16:06.641 100663296 bytes (101 MB, 96 MiB) copied, 0.524493 s, 192 MB/s 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:06.641 [2024-11-20 16:04:04.791765] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:06.641 [2024-11-20 16:04:04.825102] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:06.641 "name": "raid_bdev1", 00:16:06.641 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:06.641 "strip_size_kb": 64, 00:16:06.641 "state": "online", 00:16:06.641 "raid_level": "raid5f", 00:16:06.641 "superblock": false, 00:16:06.641 "num_base_bdevs": 4, 00:16:06.641 "num_base_bdevs_discovered": 3, 00:16:06.641 "num_base_bdevs_operational": 3, 00:16:06.641 "base_bdevs_list": [ 00:16:06.641 { 00:16:06.641 "name": null, 00:16:06.641 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:06.641 "is_configured": false, 00:16:06.641 "data_offset": 0, 00:16:06.641 "data_size": 65536 00:16:06.641 }, 00:16:06.641 { 00:16:06.641 "name": "BaseBdev2", 00:16:06.641 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:06.641 "is_configured": true, 00:16:06.641 "data_offset": 0, 00:16:06.641 "data_size": 65536 00:16:06.641 }, 00:16:06.641 { 00:16:06.641 "name": "BaseBdev3", 00:16:06.641 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:06.641 "is_configured": true, 00:16:06.641 "data_offset": 0, 00:16:06.641 "data_size": 65536 00:16:06.641 }, 00:16:06.641 { 00:16:06.641 "name": "BaseBdev4", 00:16:06.641 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:06.641 "is_configured": true, 00:16:06.641 "data_offset": 0, 00:16:06.641 "data_size": 65536 00:16:06.641 } 00:16:06.641 ] 00:16:06.641 }' 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:06.641 16:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:06.898 16:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:06.898 16:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.898 16:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:06.898 [2024-11-20 16:04:05.129176] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:06.898 [2024-11-20 16:04:05.139416] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b750 00:16:06.898 16:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.898 16:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:06.898 [2024-11-20 16:04:05.146258] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:08.267 "name": "raid_bdev1", 00:16:08.267 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:08.267 "strip_size_kb": 64, 00:16:08.267 "state": "online", 00:16:08.267 "raid_level": "raid5f", 00:16:08.267 "superblock": false, 00:16:08.267 "num_base_bdevs": 4, 00:16:08.267 "num_base_bdevs_discovered": 4, 00:16:08.267 "num_base_bdevs_operational": 4, 00:16:08.267 "process": { 00:16:08.267 "type": "rebuild", 00:16:08.267 "target": "spare", 00:16:08.267 "progress": { 00:16:08.267 "blocks": 17280, 00:16:08.267 "percent": 8 00:16:08.267 } 00:16:08.267 }, 00:16:08.267 "base_bdevs_list": [ 00:16:08.267 { 00:16:08.267 "name": "spare", 00:16:08.267 "uuid": "31ef0dc3-a3bb-5dc5-9170-d97e4b514921", 00:16:08.267 "is_configured": true, 00:16:08.267 "data_offset": 0, 00:16:08.267 "data_size": 65536 00:16:08.267 }, 00:16:08.267 { 00:16:08.267 "name": "BaseBdev2", 00:16:08.267 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:08.267 "is_configured": true, 00:16:08.267 "data_offset": 0, 00:16:08.267 "data_size": 65536 00:16:08.267 }, 00:16:08.267 { 00:16:08.267 "name": "BaseBdev3", 00:16:08.267 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:08.267 "is_configured": true, 00:16:08.267 "data_offset": 0, 00:16:08.267 "data_size": 65536 00:16:08.267 }, 00:16:08.267 { 00:16:08.267 "name": "BaseBdev4", 00:16:08.267 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:08.267 "is_configured": true, 00:16:08.267 "data_offset": 0, 00:16:08.267 "data_size": 65536 00:16:08.267 } 00:16:08.267 ] 00:16:08.267 }' 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:08.267 [2024-11-20 16:04:06.247270] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:08.267 [2024-11-20 16:04:06.254914] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:08.267 [2024-11-20 16:04:06.254975] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:08.267 [2024-11-20 16:04:06.254992] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:08.267 [2024-11-20 16:04:06.255001] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.267 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:08.267 "name": "raid_bdev1", 00:16:08.267 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:08.267 "strip_size_kb": 64, 00:16:08.267 "state": "online", 00:16:08.267 "raid_level": "raid5f", 00:16:08.267 "superblock": false, 00:16:08.267 "num_base_bdevs": 4, 00:16:08.267 "num_base_bdevs_discovered": 3, 00:16:08.267 "num_base_bdevs_operational": 3, 00:16:08.267 "base_bdevs_list": [ 00:16:08.267 { 00:16:08.267 "name": null, 00:16:08.267 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:08.267 "is_configured": false, 00:16:08.267 "data_offset": 0, 00:16:08.267 "data_size": 65536 00:16:08.267 }, 00:16:08.267 { 00:16:08.267 "name": "BaseBdev2", 00:16:08.267 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:08.267 "is_configured": true, 00:16:08.267 "data_offset": 0, 00:16:08.267 "data_size": 65536 00:16:08.267 }, 00:16:08.267 { 00:16:08.267 "name": "BaseBdev3", 00:16:08.267 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:08.268 "is_configured": true, 00:16:08.268 "data_offset": 0, 00:16:08.268 "data_size": 65536 00:16:08.268 }, 00:16:08.268 { 00:16:08.268 "name": "BaseBdev4", 00:16:08.268 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:08.268 "is_configured": true, 00:16:08.268 "data_offset": 0, 00:16:08.268 "data_size": 65536 00:16:08.268 } 00:16:08.268 ] 00:16:08.268 }' 00:16:08.268 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:08.268 16:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:08.525 "name": "raid_bdev1", 00:16:08.525 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:08.525 "strip_size_kb": 64, 00:16:08.525 "state": "online", 00:16:08.525 "raid_level": "raid5f", 00:16:08.525 "superblock": false, 00:16:08.525 "num_base_bdevs": 4, 00:16:08.525 "num_base_bdevs_discovered": 3, 00:16:08.525 "num_base_bdevs_operational": 3, 00:16:08.525 "base_bdevs_list": [ 00:16:08.525 { 00:16:08.525 "name": null, 00:16:08.525 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:08.525 "is_configured": false, 00:16:08.525 "data_offset": 0, 00:16:08.525 "data_size": 65536 00:16:08.525 }, 00:16:08.525 { 00:16:08.525 "name": "BaseBdev2", 00:16:08.525 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:08.525 "is_configured": true, 00:16:08.525 "data_offset": 0, 00:16:08.525 "data_size": 65536 00:16:08.525 }, 00:16:08.525 { 00:16:08.525 "name": "BaseBdev3", 00:16:08.525 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:08.525 "is_configured": true, 00:16:08.525 "data_offset": 0, 00:16:08.525 "data_size": 65536 00:16:08.525 }, 00:16:08.525 { 00:16:08.525 "name": "BaseBdev4", 00:16:08.525 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:08.525 "is_configured": true, 00:16:08.525 "data_offset": 0, 00:16:08.525 "data_size": 65536 00:16:08.525 } 00:16:08.525 ] 00:16:08.525 }' 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:08.525 [2024-11-20 16:04:06.682112] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:08.525 [2024-11-20 16:04:06.691771] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b820 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.525 16:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:08.525 [2024-11-20 16:04:06.698094] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:09.457 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:09.457 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:09.457 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:09.457 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:09.457 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:09.457 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:09.457 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:09.457 16:04:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:09.457 16:04:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:09.717 "name": "raid_bdev1", 00:16:09.717 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:09.717 "strip_size_kb": 64, 00:16:09.717 "state": "online", 00:16:09.717 "raid_level": "raid5f", 00:16:09.717 "superblock": false, 00:16:09.717 "num_base_bdevs": 4, 00:16:09.717 "num_base_bdevs_discovered": 4, 00:16:09.717 "num_base_bdevs_operational": 4, 00:16:09.717 "process": { 00:16:09.717 "type": "rebuild", 00:16:09.717 "target": "spare", 00:16:09.717 "progress": { 00:16:09.717 "blocks": 19200, 00:16:09.717 "percent": 9 00:16:09.717 } 00:16:09.717 }, 00:16:09.717 "base_bdevs_list": [ 00:16:09.717 { 00:16:09.717 "name": "spare", 00:16:09.717 "uuid": "31ef0dc3-a3bb-5dc5-9170-d97e4b514921", 00:16:09.717 "is_configured": true, 00:16:09.717 "data_offset": 0, 00:16:09.717 "data_size": 65536 00:16:09.717 }, 00:16:09.717 { 00:16:09.717 "name": "BaseBdev2", 00:16:09.717 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:09.717 "is_configured": true, 00:16:09.717 "data_offset": 0, 00:16:09.717 "data_size": 65536 00:16:09.717 }, 00:16:09.717 { 00:16:09.717 "name": "BaseBdev3", 00:16:09.717 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:09.717 "is_configured": true, 00:16:09.717 "data_offset": 0, 00:16:09.717 "data_size": 65536 00:16:09.717 }, 00:16:09.717 { 00:16:09.717 "name": "BaseBdev4", 00:16:09.717 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:09.717 "is_configured": true, 00:16:09.717 "data_offset": 0, 00:16:09.717 "data_size": 65536 00:16:09.717 } 00:16:09.717 ] 00:16:09.717 }' 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=501 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:09.717 16:04:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:09.718 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:09.718 16:04:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:09.718 16:04:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:09.718 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:09.718 "name": "raid_bdev1", 00:16:09.718 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:09.718 "strip_size_kb": 64, 00:16:09.718 "state": "online", 00:16:09.718 "raid_level": "raid5f", 00:16:09.718 "superblock": false, 00:16:09.718 "num_base_bdevs": 4, 00:16:09.718 "num_base_bdevs_discovered": 4, 00:16:09.718 "num_base_bdevs_operational": 4, 00:16:09.718 "process": { 00:16:09.718 "type": "rebuild", 00:16:09.718 "target": "spare", 00:16:09.718 "progress": { 00:16:09.718 "blocks": 21120, 00:16:09.718 "percent": 10 00:16:09.718 } 00:16:09.718 }, 00:16:09.718 "base_bdevs_list": [ 00:16:09.718 { 00:16:09.718 "name": "spare", 00:16:09.718 "uuid": "31ef0dc3-a3bb-5dc5-9170-d97e4b514921", 00:16:09.718 "is_configured": true, 00:16:09.718 "data_offset": 0, 00:16:09.718 "data_size": 65536 00:16:09.718 }, 00:16:09.718 { 00:16:09.718 "name": "BaseBdev2", 00:16:09.718 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:09.718 "is_configured": true, 00:16:09.718 "data_offset": 0, 00:16:09.718 "data_size": 65536 00:16:09.718 }, 00:16:09.718 { 00:16:09.718 "name": "BaseBdev3", 00:16:09.718 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:09.718 "is_configured": true, 00:16:09.718 "data_offset": 0, 00:16:09.718 "data_size": 65536 00:16:09.718 }, 00:16:09.718 { 00:16:09.718 "name": "BaseBdev4", 00:16:09.718 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:09.718 "is_configured": true, 00:16:09.718 "data_offset": 0, 00:16:09.718 "data_size": 65536 00:16:09.718 } 00:16:09.718 ] 00:16:09.718 }' 00:16:09.718 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:09.718 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:09.718 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:09.718 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:09.718 16:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:10.696 16:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:10.696 16:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:10.696 16:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:10.696 16:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:10.696 16:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:10.696 16:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:10.696 16:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:10.696 16:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:10.696 16:04:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:10.696 16:04:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:10.696 16:04:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:10.696 16:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:10.696 "name": "raid_bdev1", 00:16:10.696 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:10.696 "strip_size_kb": 64, 00:16:10.696 "state": "online", 00:16:10.696 "raid_level": "raid5f", 00:16:10.696 "superblock": false, 00:16:10.696 "num_base_bdevs": 4, 00:16:10.696 "num_base_bdevs_discovered": 4, 00:16:10.696 "num_base_bdevs_operational": 4, 00:16:10.696 "process": { 00:16:10.696 "type": "rebuild", 00:16:10.696 "target": "spare", 00:16:10.696 "progress": { 00:16:10.696 "blocks": 40320, 00:16:10.696 "percent": 20 00:16:10.696 } 00:16:10.696 }, 00:16:10.696 "base_bdevs_list": [ 00:16:10.696 { 00:16:10.696 "name": "spare", 00:16:10.696 "uuid": "31ef0dc3-a3bb-5dc5-9170-d97e4b514921", 00:16:10.696 "is_configured": true, 00:16:10.696 "data_offset": 0, 00:16:10.696 "data_size": 65536 00:16:10.696 }, 00:16:10.696 { 00:16:10.696 "name": "BaseBdev2", 00:16:10.696 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:10.696 "is_configured": true, 00:16:10.696 "data_offset": 0, 00:16:10.696 "data_size": 65536 00:16:10.696 }, 00:16:10.696 { 00:16:10.696 "name": "BaseBdev3", 00:16:10.696 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:10.696 "is_configured": true, 00:16:10.696 "data_offset": 0, 00:16:10.696 "data_size": 65536 00:16:10.696 }, 00:16:10.696 { 00:16:10.696 "name": "BaseBdev4", 00:16:10.696 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:10.696 "is_configured": true, 00:16:10.696 "data_offset": 0, 00:16:10.696 "data_size": 65536 00:16:10.696 } 00:16:10.696 ] 00:16:10.696 }' 00:16:10.696 16:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:10.954 16:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:10.954 16:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:10.954 16:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:10.954 16:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:11.887 16:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:11.887 16:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:11.887 16:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:11.887 16:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:11.887 16:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:11.887 16:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:11.887 16:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:11.887 16:04:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:11.887 16:04:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:11.887 16:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:11.887 16:04:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:11.887 16:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:11.887 "name": "raid_bdev1", 00:16:11.887 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:11.887 "strip_size_kb": 64, 00:16:11.887 "state": "online", 00:16:11.887 "raid_level": "raid5f", 00:16:11.887 "superblock": false, 00:16:11.887 "num_base_bdevs": 4, 00:16:11.887 "num_base_bdevs_discovered": 4, 00:16:11.887 "num_base_bdevs_operational": 4, 00:16:11.887 "process": { 00:16:11.887 "type": "rebuild", 00:16:11.887 "target": "spare", 00:16:11.887 "progress": { 00:16:11.887 "blocks": 61440, 00:16:11.887 "percent": 31 00:16:11.887 } 00:16:11.887 }, 00:16:11.887 "base_bdevs_list": [ 00:16:11.887 { 00:16:11.887 "name": "spare", 00:16:11.887 "uuid": "31ef0dc3-a3bb-5dc5-9170-d97e4b514921", 00:16:11.887 "is_configured": true, 00:16:11.887 "data_offset": 0, 00:16:11.887 "data_size": 65536 00:16:11.887 }, 00:16:11.887 { 00:16:11.887 "name": "BaseBdev2", 00:16:11.887 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:11.887 "is_configured": true, 00:16:11.887 "data_offset": 0, 00:16:11.887 "data_size": 65536 00:16:11.887 }, 00:16:11.887 { 00:16:11.887 "name": "BaseBdev3", 00:16:11.887 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:11.887 "is_configured": true, 00:16:11.887 "data_offset": 0, 00:16:11.887 "data_size": 65536 00:16:11.887 }, 00:16:11.887 { 00:16:11.887 "name": "BaseBdev4", 00:16:11.887 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:11.887 "is_configured": true, 00:16:11.887 "data_offset": 0, 00:16:11.887 "data_size": 65536 00:16:11.887 } 00:16:11.887 ] 00:16:11.887 }' 00:16:11.887 16:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:11.887 16:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:11.887 16:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:11.887 16:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:11.887 16:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:13.261 16:04:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:13.261 16:04:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:13.261 16:04:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:13.261 16:04:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:13.261 16:04:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:13.261 16:04:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:13.261 16:04:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:13.261 16:04:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:13.261 16:04:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:13.261 16:04:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:13.262 16:04:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:13.262 16:04:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:13.262 "name": "raid_bdev1", 00:16:13.262 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:13.262 "strip_size_kb": 64, 00:16:13.262 "state": "online", 00:16:13.262 "raid_level": "raid5f", 00:16:13.262 "superblock": false, 00:16:13.262 "num_base_bdevs": 4, 00:16:13.262 "num_base_bdevs_discovered": 4, 00:16:13.262 "num_base_bdevs_operational": 4, 00:16:13.262 "process": { 00:16:13.262 "type": "rebuild", 00:16:13.262 "target": "spare", 00:16:13.262 "progress": { 00:16:13.262 "blocks": 82560, 00:16:13.262 "percent": 41 00:16:13.262 } 00:16:13.262 }, 00:16:13.262 "base_bdevs_list": [ 00:16:13.262 { 00:16:13.262 "name": "spare", 00:16:13.262 "uuid": "31ef0dc3-a3bb-5dc5-9170-d97e4b514921", 00:16:13.262 "is_configured": true, 00:16:13.262 "data_offset": 0, 00:16:13.262 "data_size": 65536 00:16:13.262 }, 00:16:13.262 { 00:16:13.262 "name": "BaseBdev2", 00:16:13.262 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:13.262 "is_configured": true, 00:16:13.262 "data_offset": 0, 00:16:13.262 "data_size": 65536 00:16:13.262 }, 00:16:13.262 { 00:16:13.262 "name": "BaseBdev3", 00:16:13.262 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:13.262 "is_configured": true, 00:16:13.262 "data_offset": 0, 00:16:13.262 "data_size": 65536 00:16:13.262 }, 00:16:13.262 { 00:16:13.262 "name": "BaseBdev4", 00:16:13.262 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:13.262 "is_configured": true, 00:16:13.262 "data_offset": 0, 00:16:13.262 "data_size": 65536 00:16:13.262 } 00:16:13.262 ] 00:16:13.262 }' 00:16:13.262 16:04:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:13.262 16:04:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:13.262 16:04:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:13.262 16:04:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:13.262 16:04:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:14.195 16:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:14.195 16:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:14.195 16:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:14.195 16:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:14.195 16:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:14.195 16:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:14.195 16:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:14.195 16:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:14.195 16:04:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:14.195 16:04:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:14.195 16:04:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:14.195 16:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:14.195 "name": "raid_bdev1", 00:16:14.195 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:14.195 "strip_size_kb": 64, 00:16:14.195 "state": "online", 00:16:14.195 "raid_level": "raid5f", 00:16:14.195 "superblock": false, 00:16:14.195 "num_base_bdevs": 4, 00:16:14.195 "num_base_bdevs_discovered": 4, 00:16:14.195 "num_base_bdevs_operational": 4, 00:16:14.195 "process": { 00:16:14.195 "type": "rebuild", 00:16:14.195 "target": "spare", 00:16:14.195 "progress": { 00:16:14.195 "blocks": 103680, 00:16:14.195 "percent": 52 00:16:14.195 } 00:16:14.195 }, 00:16:14.195 "base_bdevs_list": [ 00:16:14.195 { 00:16:14.195 "name": "spare", 00:16:14.195 "uuid": "31ef0dc3-a3bb-5dc5-9170-d97e4b514921", 00:16:14.195 "is_configured": true, 00:16:14.195 "data_offset": 0, 00:16:14.195 "data_size": 65536 00:16:14.195 }, 00:16:14.195 { 00:16:14.195 "name": "BaseBdev2", 00:16:14.195 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:14.195 "is_configured": true, 00:16:14.195 "data_offset": 0, 00:16:14.195 "data_size": 65536 00:16:14.195 }, 00:16:14.195 { 00:16:14.195 "name": "BaseBdev3", 00:16:14.195 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:14.195 "is_configured": true, 00:16:14.195 "data_offset": 0, 00:16:14.195 "data_size": 65536 00:16:14.195 }, 00:16:14.195 { 00:16:14.195 "name": "BaseBdev4", 00:16:14.195 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:14.195 "is_configured": true, 00:16:14.195 "data_offset": 0, 00:16:14.195 "data_size": 65536 00:16:14.195 } 00:16:14.195 ] 00:16:14.195 }' 00:16:14.195 16:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:14.195 16:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:14.195 16:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:14.195 16:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:14.195 16:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:15.129 16:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:15.129 16:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:15.129 16:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:15.129 16:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:15.129 16:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:15.129 16:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:15.129 16:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:15.129 16:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:15.129 16:04:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:15.129 16:04:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:15.129 16:04:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:15.129 16:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:15.129 "name": "raid_bdev1", 00:16:15.129 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:15.129 "strip_size_kb": 64, 00:16:15.129 "state": "online", 00:16:15.129 "raid_level": "raid5f", 00:16:15.129 "superblock": false, 00:16:15.129 "num_base_bdevs": 4, 00:16:15.129 "num_base_bdevs_discovered": 4, 00:16:15.129 "num_base_bdevs_operational": 4, 00:16:15.129 "process": { 00:16:15.129 "type": "rebuild", 00:16:15.129 "target": "spare", 00:16:15.129 "progress": { 00:16:15.129 "blocks": 124800, 00:16:15.129 "percent": 63 00:16:15.129 } 00:16:15.129 }, 00:16:15.129 "base_bdevs_list": [ 00:16:15.129 { 00:16:15.129 "name": "spare", 00:16:15.129 "uuid": "31ef0dc3-a3bb-5dc5-9170-d97e4b514921", 00:16:15.129 "is_configured": true, 00:16:15.129 "data_offset": 0, 00:16:15.129 "data_size": 65536 00:16:15.129 }, 00:16:15.129 { 00:16:15.129 "name": "BaseBdev2", 00:16:15.129 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:15.129 "is_configured": true, 00:16:15.129 "data_offset": 0, 00:16:15.129 "data_size": 65536 00:16:15.129 }, 00:16:15.129 { 00:16:15.129 "name": "BaseBdev3", 00:16:15.129 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:15.129 "is_configured": true, 00:16:15.129 "data_offset": 0, 00:16:15.129 "data_size": 65536 00:16:15.129 }, 00:16:15.129 { 00:16:15.129 "name": "BaseBdev4", 00:16:15.129 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:15.129 "is_configured": true, 00:16:15.129 "data_offset": 0, 00:16:15.129 "data_size": 65536 00:16:15.129 } 00:16:15.129 ] 00:16:15.129 }' 00:16:15.129 16:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:15.387 16:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:15.387 16:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:15.387 16:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:15.387 16:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:16.320 16:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:16.320 16:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:16.320 16:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:16.320 16:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:16.320 16:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:16.321 16:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:16.321 16:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:16.321 16:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:16.321 16:04:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.321 16:04:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:16.321 16:04:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.321 16:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:16.321 "name": "raid_bdev1", 00:16:16.321 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:16.321 "strip_size_kb": 64, 00:16:16.321 "state": "online", 00:16:16.321 "raid_level": "raid5f", 00:16:16.321 "superblock": false, 00:16:16.321 "num_base_bdevs": 4, 00:16:16.321 "num_base_bdevs_discovered": 4, 00:16:16.321 "num_base_bdevs_operational": 4, 00:16:16.321 "process": { 00:16:16.321 "type": "rebuild", 00:16:16.321 "target": "spare", 00:16:16.321 "progress": { 00:16:16.321 "blocks": 145920, 00:16:16.321 "percent": 74 00:16:16.321 } 00:16:16.321 }, 00:16:16.321 "base_bdevs_list": [ 00:16:16.321 { 00:16:16.321 "name": "spare", 00:16:16.321 "uuid": "31ef0dc3-a3bb-5dc5-9170-d97e4b514921", 00:16:16.321 "is_configured": true, 00:16:16.321 "data_offset": 0, 00:16:16.321 "data_size": 65536 00:16:16.321 }, 00:16:16.321 { 00:16:16.321 "name": "BaseBdev2", 00:16:16.321 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:16.321 "is_configured": true, 00:16:16.321 "data_offset": 0, 00:16:16.321 "data_size": 65536 00:16:16.321 }, 00:16:16.321 { 00:16:16.321 "name": "BaseBdev3", 00:16:16.321 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:16.321 "is_configured": true, 00:16:16.321 "data_offset": 0, 00:16:16.321 "data_size": 65536 00:16:16.321 }, 00:16:16.321 { 00:16:16.321 "name": "BaseBdev4", 00:16:16.321 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:16.321 "is_configured": true, 00:16:16.321 "data_offset": 0, 00:16:16.321 "data_size": 65536 00:16:16.321 } 00:16:16.321 ] 00:16:16.321 }' 00:16:16.321 16:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:16.321 16:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:16.321 16:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:16.321 16:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:16.321 16:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:17.693 16:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:17.693 16:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:17.693 16:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:17.693 16:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:17.693 16:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:17.693 16:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:17.693 16:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:17.693 16:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:17.693 16:04:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:17.693 16:04:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:17.693 16:04:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:17.693 16:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:17.693 "name": "raid_bdev1", 00:16:17.693 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:17.693 "strip_size_kb": 64, 00:16:17.693 "state": "online", 00:16:17.693 "raid_level": "raid5f", 00:16:17.693 "superblock": false, 00:16:17.693 "num_base_bdevs": 4, 00:16:17.693 "num_base_bdevs_discovered": 4, 00:16:17.693 "num_base_bdevs_operational": 4, 00:16:17.693 "process": { 00:16:17.693 "type": "rebuild", 00:16:17.693 "target": "spare", 00:16:17.693 "progress": { 00:16:17.693 "blocks": 167040, 00:16:17.693 "percent": 84 00:16:17.693 } 00:16:17.693 }, 00:16:17.693 "base_bdevs_list": [ 00:16:17.693 { 00:16:17.693 "name": "spare", 00:16:17.693 "uuid": "31ef0dc3-a3bb-5dc5-9170-d97e4b514921", 00:16:17.693 "is_configured": true, 00:16:17.693 "data_offset": 0, 00:16:17.693 "data_size": 65536 00:16:17.693 }, 00:16:17.693 { 00:16:17.693 "name": "BaseBdev2", 00:16:17.693 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:17.693 "is_configured": true, 00:16:17.693 "data_offset": 0, 00:16:17.693 "data_size": 65536 00:16:17.693 }, 00:16:17.693 { 00:16:17.693 "name": "BaseBdev3", 00:16:17.693 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:17.693 "is_configured": true, 00:16:17.693 "data_offset": 0, 00:16:17.693 "data_size": 65536 00:16:17.693 }, 00:16:17.693 { 00:16:17.693 "name": "BaseBdev4", 00:16:17.693 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:17.693 "is_configured": true, 00:16:17.693 "data_offset": 0, 00:16:17.693 "data_size": 65536 00:16:17.693 } 00:16:17.693 ] 00:16:17.693 }' 00:16:17.693 16:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:17.693 16:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:17.693 16:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:17.693 16:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:17.693 16:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:18.686 16:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:18.686 16:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:18.686 16:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:18.686 16:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:18.686 16:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:18.686 16:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:18.686 16:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:18.686 16:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:18.686 16:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.686 16:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:18.686 16:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.686 16:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:18.686 "name": "raid_bdev1", 00:16:18.686 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:18.686 "strip_size_kb": 64, 00:16:18.686 "state": "online", 00:16:18.686 "raid_level": "raid5f", 00:16:18.686 "superblock": false, 00:16:18.686 "num_base_bdevs": 4, 00:16:18.686 "num_base_bdevs_discovered": 4, 00:16:18.686 "num_base_bdevs_operational": 4, 00:16:18.686 "process": { 00:16:18.686 "type": "rebuild", 00:16:18.686 "target": "spare", 00:16:18.686 "progress": { 00:16:18.686 "blocks": 188160, 00:16:18.686 "percent": 95 00:16:18.686 } 00:16:18.686 }, 00:16:18.686 "base_bdevs_list": [ 00:16:18.686 { 00:16:18.686 "name": "spare", 00:16:18.686 "uuid": "31ef0dc3-a3bb-5dc5-9170-d97e4b514921", 00:16:18.686 "is_configured": true, 00:16:18.686 "data_offset": 0, 00:16:18.686 "data_size": 65536 00:16:18.686 }, 00:16:18.686 { 00:16:18.686 "name": "BaseBdev2", 00:16:18.686 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:18.686 "is_configured": true, 00:16:18.686 "data_offset": 0, 00:16:18.686 "data_size": 65536 00:16:18.686 }, 00:16:18.686 { 00:16:18.686 "name": "BaseBdev3", 00:16:18.686 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:18.686 "is_configured": true, 00:16:18.686 "data_offset": 0, 00:16:18.686 "data_size": 65536 00:16:18.686 }, 00:16:18.686 { 00:16:18.686 "name": "BaseBdev4", 00:16:18.686 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:18.686 "is_configured": true, 00:16:18.686 "data_offset": 0, 00:16:18.686 "data_size": 65536 00:16:18.686 } 00:16:18.686 ] 00:16:18.686 }' 00:16:18.686 16:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:18.686 16:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:18.686 16:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:18.686 16:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:18.686 16:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:18.944 [2024-11-20 16:04:17.068911] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:18.944 [2024-11-20 16:04:17.069132] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:18.944 [2024-11-20 16:04:17.069182] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:19.508 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:19.508 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:19.508 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:19.508 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:19.508 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:19.508 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:19.508 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:19.508 16:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:19.508 16:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:19.508 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:19.766 "name": "raid_bdev1", 00:16:19.766 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:19.766 "strip_size_kb": 64, 00:16:19.766 "state": "online", 00:16:19.766 "raid_level": "raid5f", 00:16:19.766 "superblock": false, 00:16:19.766 "num_base_bdevs": 4, 00:16:19.766 "num_base_bdevs_discovered": 4, 00:16:19.766 "num_base_bdevs_operational": 4, 00:16:19.766 "base_bdevs_list": [ 00:16:19.766 { 00:16:19.766 "name": "spare", 00:16:19.766 "uuid": "31ef0dc3-a3bb-5dc5-9170-d97e4b514921", 00:16:19.766 "is_configured": true, 00:16:19.766 "data_offset": 0, 00:16:19.766 "data_size": 65536 00:16:19.766 }, 00:16:19.766 { 00:16:19.766 "name": "BaseBdev2", 00:16:19.766 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:19.766 "is_configured": true, 00:16:19.766 "data_offset": 0, 00:16:19.766 "data_size": 65536 00:16:19.766 }, 00:16:19.766 { 00:16:19.766 "name": "BaseBdev3", 00:16:19.766 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:19.766 "is_configured": true, 00:16:19.766 "data_offset": 0, 00:16:19.766 "data_size": 65536 00:16:19.766 }, 00:16:19.766 { 00:16:19.766 "name": "BaseBdev4", 00:16:19.766 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:19.766 "is_configured": true, 00:16:19.766 "data_offset": 0, 00:16:19.766 "data_size": 65536 00:16:19.766 } 00:16:19.766 ] 00:16:19.766 }' 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:19.766 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:19.766 "name": "raid_bdev1", 00:16:19.766 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:19.766 "strip_size_kb": 64, 00:16:19.766 "state": "online", 00:16:19.766 "raid_level": "raid5f", 00:16:19.766 "superblock": false, 00:16:19.766 "num_base_bdevs": 4, 00:16:19.767 "num_base_bdevs_discovered": 4, 00:16:19.767 "num_base_bdevs_operational": 4, 00:16:19.767 "base_bdevs_list": [ 00:16:19.767 { 00:16:19.767 "name": "spare", 00:16:19.767 "uuid": "31ef0dc3-a3bb-5dc5-9170-d97e4b514921", 00:16:19.767 "is_configured": true, 00:16:19.767 "data_offset": 0, 00:16:19.767 "data_size": 65536 00:16:19.767 }, 00:16:19.767 { 00:16:19.767 "name": "BaseBdev2", 00:16:19.767 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:19.767 "is_configured": true, 00:16:19.767 "data_offset": 0, 00:16:19.767 "data_size": 65536 00:16:19.767 }, 00:16:19.767 { 00:16:19.767 "name": "BaseBdev3", 00:16:19.767 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:19.767 "is_configured": true, 00:16:19.767 "data_offset": 0, 00:16:19.767 "data_size": 65536 00:16:19.767 }, 00:16:19.767 { 00:16:19.767 "name": "BaseBdev4", 00:16:19.767 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:19.767 "is_configured": true, 00:16:19.767 "data_offset": 0, 00:16:19.767 "data_size": 65536 00:16:19.767 } 00:16:19.767 ] 00:16:19.767 }' 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:19.767 "name": "raid_bdev1", 00:16:19.767 "uuid": "079627fe-37cc-4bae-acca-9e16fb89fbe1", 00:16:19.767 "strip_size_kb": 64, 00:16:19.767 "state": "online", 00:16:19.767 "raid_level": "raid5f", 00:16:19.767 "superblock": false, 00:16:19.767 "num_base_bdevs": 4, 00:16:19.767 "num_base_bdevs_discovered": 4, 00:16:19.767 "num_base_bdevs_operational": 4, 00:16:19.767 "base_bdevs_list": [ 00:16:19.767 { 00:16:19.767 "name": "spare", 00:16:19.767 "uuid": "31ef0dc3-a3bb-5dc5-9170-d97e4b514921", 00:16:19.767 "is_configured": true, 00:16:19.767 "data_offset": 0, 00:16:19.767 "data_size": 65536 00:16:19.767 }, 00:16:19.767 { 00:16:19.767 "name": "BaseBdev2", 00:16:19.767 "uuid": "5f4bdfd0-ab31-5ce6-8fbb-df015efb2ebc", 00:16:19.767 "is_configured": true, 00:16:19.767 "data_offset": 0, 00:16:19.767 "data_size": 65536 00:16:19.767 }, 00:16:19.767 { 00:16:19.767 "name": "BaseBdev3", 00:16:19.767 "uuid": "3ccb1a1a-8082-5c9c-adee-2ee3f142e806", 00:16:19.767 "is_configured": true, 00:16:19.767 "data_offset": 0, 00:16:19.767 "data_size": 65536 00:16:19.767 }, 00:16:19.767 { 00:16:19.767 "name": "BaseBdev4", 00:16:19.767 "uuid": "aea56f84-33b3-5844-b5f8-c5774d385774", 00:16:19.767 "is_configured": true, 00:16:19.767 "data_offset": 0, 00:16:19.767 "data_size": 65536 00:16:19.767 } 00:16:19.767 ] 00:16:19.767 }' 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:19.767 16:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:20.025 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:20.025 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.025 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:20.025 [2024-11-20 16:04:18.245931] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:20.025 [2024-11-20 16:04:18.245957] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:20.025 [2024-11-20 16:04:18.246019] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:20.025 [2024-11-20 16:04:18.246102] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:20.025 [2024-11-20 16:04:18.246111] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:16:20.025 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.025 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.025 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:16:20.025 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.025 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:20.025 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:16:20.283 /dev/nbd0 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:20.283 1+0 records in 00:16:20.283 1+0 records out 00:16:20.283 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000255033 s, 16.1 MB/s 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:16:20.283 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:16:20.541 /dev/nbd1 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:20.541 1+0 records in 00:16:20.541 1+0 records out 00:16:20.541 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000248965 s, 16.5 MB/s 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:20.541 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:16:20.798 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:16:20.798 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:20.798 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:20.798 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:20.798 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:16:20.798 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:20.798 16:04:18 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:21.057 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:21.057 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:21.057 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:21.057 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:21.057 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:21.057 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:21.057 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:16:21.057 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:16:21.057 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:21.057 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:16:21.316 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:16:21.316 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:16:21.316 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:16:21.316 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:21.316 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:21.316 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:16:21.316 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:16:21.316 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:16:21.316 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:16:21.316 16:04:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 82224 00:16:21.316 16:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 82224 ']' 00:16:21.316 16:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 82224 00:16:21.316 16:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:16:21.316 16:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:21.316 16:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 82224 00:16:21.316 killing process with pid 82224 00:16:21.316 Received shutdown signal, test time was about 60.000000 seconds 00:16:21.316 00:16:21.316 Latency(us) 00:16:21.316 [2024-11-20T16:04:19.568Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:21.317 [2024-11-20T16:04:19.568Z] =================================================================================================================== 00:16:21.317 [2024-11-20T16:04:19.568Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:21.317 16:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:21.317 16:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:21.317 16:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 82224' 00:16:21.317 16:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@973 -- # kill 82224 00:16:21.317 [2024-11-20 16:04:19.343266] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:21.317 16:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@978 -- # wait 82224 00:16:21.575 [2024-11-20 16:04:19.586458] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:16:22.189 00:16:22.189 real 0m17.881s 00:16:22.189 user 0m20.970s 00:16:22.189 sys 0m1.753s 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:22.189 ************************************ 00:16:22.189 END TEST raid5f_rebuild_test 00:16:22.189 ************************************ 00:16:22.189 16:04:20 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 4 true false true 00:16:22.189 16:04:20 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:16:22.189 16:04:20 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:22.189 16:04:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:22.189 ************************************ 00:16:22.189 START TEST raid5f_rebuild_test_sb 00:16:22.189 ************************************ 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 4 true false true 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=82724 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 82724 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 82724 ']' 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:22.189 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:22.189 16:04:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:22.189 [2024-11-20 16:04:20.266754] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:16:22.189 [2024-11-20 16:04:20.267032] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid82724 ] 00:16:22.189 I/O size of 3145728 is greater than zero copy threshold (65536). 00:16:22.189 Zero copy mechanism will not be used. 00:16:22.447 [2024-11-20 16:04:20.419968] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:22.447 [2024-11-20 16:04:20.506491] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:22.447 [2024-11-20 16:04:20.619685] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:22.447 [2024-11-20 16:04:20.619863] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.014 BaseBdev1_malloc 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.014 [2024-11-20 16:04:21.140910] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:23.014 [2024-11-20 16:04:21.140964] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:23.014 [2024-11-20 16:04:21.140982] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:23.014 [2024-11-20 16:04:21.140992] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:23.014 [2024-11-20 16:04:21.142838] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:23.014 [2024-11-20 16:04:21.142962] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:23.014 BaseBdev1 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.014 BaseBdev2_malloc 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.014 [2024-11-20 16:04:21.172948] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:16:23.014 [2024-11-20 16:04:21.172997] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:23.014 [2024-11-20 16:04:21.173014] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:23.014 [2024-11-20 16:04:21.173022] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:23.014 [2024-11-20 16:04:21.174806] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:23.014 [2024-11-20 16:04:21.174835] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:23.014 BaseBdev2 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.014 BaseBdev3_malloc 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.014 [2024-11-20 16:04:21.219015] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:16:23.014 [2024-11-20 16:04:21.219067] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:23.014 [2024-11-20 16:04:21.219086] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:16:23.014 [2024-11-20 16:04:21.219097] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:23.014 [2024-11-20 16:04:21.220921] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:23.014 [2024-11-20 16:04:21.220954] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:16:23.014 BaseBdev3 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:23.014 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:16:23.015 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.015 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.015 BaseBdev4_malloc 00:16:23.015 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.015 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:16:23.015 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.015 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.015 [2024-11-20 16:04:21.251263] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:16:23.015 [2024-11-20 16:04:21.251315] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:23.015 [2024-11-20 16:04:21.251331] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:16:23.015 [2024-11-20 16:04:21.251341] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:23.015 [2024-11-20 16:04:21.253141] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:23.015 [2024-11-20 16:04:21.253176] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:16:23.015 BaseBdev4 00:16:23.015 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.015 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:16:23.015 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.015 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.273 spare_malloc 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.273 spare_delay 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.273 [2024-11-20 16:04:21.291128] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:23.273 [2024-11-20 16:04:21.291174] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:23.273 [2024-11-20 16:04:21.291190] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:16:23.273 [2024-11-20 16:04:21.291199] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:23.273 [2024-11-20 16:04:21.292980] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:23.273 [2024-11-20 16:04:21.293011] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:23.273 spare 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.273 [2024-11-20 16:04:21.299171] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:23.273 [2024-11-20 16:04:21.300700] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:23.273 [2024-11-20 16:04:21.300749] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:16:23.273 [2024-11-20 16:04:21.300792] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:16:23.273 [2024-11-20 16:04:21.300939] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:16:23.273 [2024-11-20 16:04:21.300951] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:16:23.273 [2024-11-20 16:04:21.301163] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:16:23.273 [2024-11-20 16:04:21.305248] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:16:23.273 [2024-11-20 16:04:21.305277] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:16:23.273 [2024-11-20 16:04:21.305442] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:23.273 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:23.274 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:23.274 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:23.274 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:23.274 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:23.274 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.274 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.274 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.274 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:23.274 "name": "raid_bdev1", 00:16:23.274 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:23.274 "strip_size_kb": 64, 00:16:23.274 "state": "online", 00:16:23.274 "raid_level": "raid5f", 00:16:23.274 "superblock": true, 00:16:23.274 "num_base_bdevs": 4, 00:16:23.274 "num_base_bdevs_discovered": 4, 00:16:23.274 "num_base_bdevs_operational": 4, 00:16:23.274 "base_bdevs_list": [ 00:16:23.274 { 00:16:23.274 "name": "BaseBdev1", 00:16:23.274 "uuid": "c3a70646-ac44-509e-b84f-0fe0eabc966d", 00:16:23.274 "is_configured": true, 00:16:23.274 "data_offset": 2048, 00:16:23.274 "data_size": 63488 00:16:23.274 }, 00:16:23.274 { 00:16:23.274 "name": "BaseBdev2", 00:16:23.274 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:23.274 "is_configured": true, 00:16:23.274 "data_offset": 2048, 00:16:23.274 "data_size": 63488 00:16:23.274 }, 00:16:23.274 { 00:16:23.274 "name": "BaseBdev3", 00:16:23.274 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:23.274 "is_configured": true, 00:16:23.274 "data_offset": 2048, 00:16:23.274 "data_size": 63488 00:16:23.274 }, 00:16:23.274 { 00:16:23.274 "name": "BaseBdev4", 00:16:23.274 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:23.274 "is_configured": true, 00:16:23.274 "data_offset": 2048, 00:16:23.274 "data_size": 63488 00:16:23.274 } 00:16:23.274 ] 00:16:23.274 }' 00:16:23.274 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:23.274 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.532 [2024-11-20 16:04:21.626144] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=190464 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:23.532 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:16:23.790 [2024-11-20 16:04:21.874052] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:16:23.790 /dev/nbd0 00:16:23.790 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:23.790 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:23.790 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:23.790 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:16:23.790 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:23.790 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:23.790 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:23.790 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:16:23.790 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:23.790 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:23.790 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:23.790 1+0 records in 00:16:23.790 1+0 records out 00:16:23.790 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000298727 s, 13.7 MB/s 00:16:23.790 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:23.790 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:16:23.790 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:23.790 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:23.790 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:16:23.790 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:23.791 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:23.791 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:16:23.791 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:16:23.791 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 192 00:16:23.791 16:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=496 oflag=direct 00:16:24.356 496+0 records in 00:16:24.356 496+0 records out 00:16:24.356 97517568 bytes (98 MB, 93 MiB) copied, 0.487583 s, 200 MB/s 00:16:24.356 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:16:24.356 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:24.356 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:24.356 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:24.356 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:16:24.356 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:24.356 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:24.613 [2024-11-20 16:04:22.632573] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:24.613 [2024-11-20 16:04:22.641376] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:24.613 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:24.614 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:24.614 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:24.614 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:24.614 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:24.614 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:24.614 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:24.614 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:24.614 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:24.614 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:24.614 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:24.614 "name": "raid_bdev1", 00:16:24.614 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:24.614 "strip_size_kb": 64, 00:16:24.614 "state": "online", 00:16:24.614 "raid_level": "raid5f", 00:16:24.614 "superblock": true, 00:16:24.614 "num_base_bdevs": 4, 00:16:24.614 "num_base_bdevs_discovered": 3, 00:16:24.614 "num_base_bdevs_operational": 3, 00:16:24.614 "base_bdevs_list": [ 00:16:24.614 { 00:16:24.614 "name": null, 00:16:24.614 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:24.614 "is_configured": false, 00:16:24.614 "data_offset": 0, 00:16:24.614 "data_size": 63488 00:16:24.614 }, 00:16:24.614 { 00:16:24.614 "name": "BaseBdev2", 00:16:24.614 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:24.614 "is_configured": true, 00:16:24.614 "data_offset": 2048, 00:16:24.614 "data_size": 63488 00:16:24.614 }, 00:16:24.614 { 00:16:24.614 "name": "BaseBdev3", 00:16:24.614 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:24.614 "is_configured": true, 00:16:24.614 "data_offset": 2048, 00:16:24.614 "data_size": 63488 00:16:24.614 }, 00:16:24.614 { 00:16:24.614 "name": "BaseBdev4", 00:16:24.614 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:24.614 "is_configured": true, 00:16:24.614 "data_offset": 2048, 00:16:24.614 "data_size": 63488 00:16:24.614 } 00:16:24.614 ] 00:16:24.614 }' 00:16:24.614 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:24.614 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:24.871 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:24.871 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:24.871 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:24.871 [2024-11-20 16:04:22.929446] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:24.871 [2024-11-20 16:04:22.938538] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002aa50 00:16:24.871 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:24.871 16:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:24.871 [2024-11-20 16:04:22.944900] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:25.805 16:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:25.805 16:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:25.805 16:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:25.805 16:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:25.805 16:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:25.805 16:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:25.805 16:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:25.805 16:04:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:25.805 16:04:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:25.805 16:04:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:25.805 16:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:25.805 "name": "raid_bdev1", 00:16:25.805 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:25.805 "strip_size_kb": 64, 00:16:25.805 "state": "online", 00:16:25.805 "raid_level": "raid5f", 00:16:25.805 "superblock": true, 00:16:25.805 "num_base_bdevs": 4, 00:16:25.805 "num_base_bdevs_discovered": 4, 00:16:25.805 "num_base_bdevs_operational": 4, 00:16:25.805 "process": { 00:16:25.805 "type": "rebuild", 00:16:25.805 "target": "spare", 00:16:25.805 "progress": { 00:16:25.805 "blocks": 17280, 00:16:25.805 "percent": 9 00:16:25.805 } 00:16:25.805 }, 00:16:25.805 "base_bdevs_list": [ 00:16:25.805 { 00:16:25.805 "name": "spare", 00:16:25.805 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:25.805 "is_configured": true, 00:16:25.805 "data_offset": 2048, 00:16:25.805 "data_size": 63488 00:16:25.805 }, 00:16:25.805 { 00:16:25.805 "name": "BaseBdev2", 00:16:25.805 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:25.805 "is_configured": true, 00:16:25.805 "data_offset": 2048, 00:16:25.805 "data_size": 63488 00:16:25.805 }, 00:16:25.805 { 00:16:25.805 "name": "BaseBdev3", 00:16:25.805 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:25.805 "is_configured": true, 00:16:25.805 "data_offset": 2048, 00:16:25.805 "data_size": 63488 00:16:25.805 }, 00:16:25.805 { 00:16:25.805 "name": "BaseBdev4", 00:16:25.805 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:25.805 "is_configured": true, 00:16:25.805 "data_offset": 2048, 00:16:25.805 "data_size": 63488 00:16:25.805 } 00:16:25.805 ] 00:16:25.805 }' 00:16:25.805 16:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:25.805 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:25.805 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:25.805 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:25.805 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:25.805 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:25.805 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:25.805 [2024-11-20 16:04:24.046036] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:26.064 [2024-11-20 16:04:24.052741] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:26.064 [2024-11-20 16:04:24.052794] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:26.064 [2024-11-20 16:04:24.052808] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:26.065 [2024-11-20 16:04:24.052816] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:26.065 "name": "raid_bdev1", 00:16:26.065 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:26.065 "strip_size_kb": 64, 00:16:26.065 "state": "online", 00:16:26.065 "raid_level": "raid5f", 00:16:26.065 "superblock": true, 00:16:26.065 "num_base_bdevs": 4, 00:16:26.065 "num_base_bdevs_discovered": 3, 00:16:26.065 "num_base_bdevs_operational": 3, 00:16:26.065 "base_bdevs_list": [ 00:16:26.065 { 00:16:26.065 "name": null, 00:16:26.065 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:26.065 "is_configured": false, 00:16:26.065 "data_offset": 0, 00:16:26.065 "data_size": 63488 00:16:26.065 }, 00:16:26.065 { 00:16:26.065 "name": "BaseBdev2", 00:16:26.065 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:26.065 "is_configured": true, 00:16:26.065 "data_offset": 2048, 00:16:26.065 "data_size": 63488 00:16:26.065 }, 00:16:26.065 { 00:16:26.065 "name": "BaseBdev3", 00:16:26.065 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:26.065 "is_configured": true, 00:16:26.065 "data_offset": 2048, 00:16:26.065 "data_size": 63488 00:16:26.065 }, 00:16:26.065 { 00:16:26.065 "name": "BaseBdev4", 00:16:26.065 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:26.065 "is_configured": true, 00:16:26.065 "data_offset": 2048, 00:16:26.065 "data_size": 63488 00:16:26.065 } 00:16:26.065 ] 00:16:26.065 }' 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:26.065 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:26.323 "name": "raid_bdev1", 00:16:26.323 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:26.323 "strip_size_kb": 64, 00:16:26.323 "state": "online", 00:16:26.323 "raid_level": "raid5f", 00:16:26.323 "superblock": true, 00:16:26.323 "num_base_bdevs": 4, 00:16:26.323 "num_base_bdevs_discovered": 3, 00:16:26.323 "num_base_bdevs_operational": 3, 00:16:26.323 "base_bdevs_list": [ 00:16:26.323 { 00:16:26.323 "name": null, 00:16:26.323 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:26.323 "is_configured": false, 00:16:26.323 "data_offset": 0, 00:16:26.323 "data_size": 63488 00:16:26.323 }, 00:16:26.323 { 00:16:26.323 "name": "BaseBdev2", 00:16:26.323 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:26.323 "is_configured": true, 00:16:26.323 "data_offset": 2048, 00:16:26.323 "data_size": 63488 00:16:26.323 }, 00:16:26.323 { 00:16:26.323 "name": "BaseBdev3", 00:16:26.323 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:26.323 "is_configured": true, 00:16:26.323 "data_offset": 2048, 00:16:26.323 "data_size": 63488 00:16:26.323 }, 00:16:26.323 { 00:16:26.323 "name": "BaseBdev4", 00:16:26.323 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:26.323 "is_configured": true, 00:16:26.323 "data_offset": 2048, 00:16:26.323 "data_size": 63488 00:16:26.323 } 00:16:26.323 ] 00:16:26.323 }' 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:26.323 [2024-11-20 16:04:24.473274] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:26.323 [2024-11-20 16:04:24.481023] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002ab20 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:26.323 16:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:26.323 [2024-11-20 16:04:24.486293] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:27.256 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:27.256 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:27.256 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:27.256 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:27.256 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:27.256 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:27.256 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:27.256 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:27.256 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:27.256 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:27.513 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:27.513 "name": "raid_bdev1", 00:16:27.513 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:27.513 "strip_size_kb": 64, 00:16:27.513 "state": "online", 00:16:27.513 "raid_level": "raid5f", 00:16:27.513 "superblock": true, 00:16:27.513 "num_base_bdevs": 4, 00:16:27.513 "num_base_bdevs_discovered": 4, 00:16:27.513 "num_base_bdevs_operational": 4, 00:16:27.513 "process": { 00:16:27.513 "type": "rebuild", 00:16:27.513 "target": "spare", 00:16:27.513 "progress": { 00:16:27.513 "blocks": 17280, 00:16:27.513 "percent": 9 00:16:27.513 } 00:16:27.513 }, 00:16:27.513 "base_bdevs_list": [ 00:16:27.513 { 00:16:27.513 "name": "spare", 00:16:27.513 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:27.513 "is_configured": true, 00:16:27.513 "data_offset": 2048, 00:16:27.513 "data_size": 63488 00:16:27.513 }, 00:16:27.513 { 00:16:27.513 "name": "BaseBdev2", 00:16:27.513 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:27.513 "is_configured": true, 00:16:27.513 "data_offset": 2048, 00:16:27.513 "data_size": 63488 00:16:27.513 }, 00:16:27.513 { 00:16:27.513 "name": "BaseBdev3", 00:16:27.513 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:27.513 "is_configured": true, 00:16:27.513 "data_offset": 2048, 00:16:27.513 "data_size": 63488 00:16:27.513 }, 00:16:27.513 { 00:16:27.513 "name": "BaseBdev4", 00:16:27.513 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:27.514 "is_configured": true, 00:16:27.514 "data_offset": 2048, 00:16:27.514 "data_size": 63488 00:16:27.514 } 00:16:27.514 ] 00:16:27.514 }' 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:16:27.514 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=519 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:27.514 "name": "raid_bdev1", 00:16:27.514 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:27.514 "strip_size_kb": 64, 00:16:27.514 "state": "online", 00:16:27.514 "raid_level": "raid5f", 00:16:27.514 "superblock": true, 00:16:27.514 "num_base_bdevs": 4, 00:16:27.514 "num_base_bdevs_discovered": 4, 00:16:27.514 "num_base_bdevs_operational": 4, 00:16:27.514 "process": { 00:16:27.514 "type": "rebuild", 00:16:27.514 "target": "spare", 00:16:27.514 "progress": { 00:16:27.514 "blocks": 19200, 00:16:27.514 "percent": 10 00:16:27.514 } 00:16:27.514 }, 00:16:27.514 "base_bdevs_list": [ 00:16:27.514 { 00:16:27.514 "name": "spare", 00:16:27.514 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:27.514 "is_configured": true, 00:16:27.514 "data_offset": 2048, 00:16:27.514 "data_size": 63488 00:16:27.514 }, 00:16:27.514 { 00:16:27.514 "name": "BaseBdev2", 00:16:27.514 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:27.514 "is_configured": true, 00:16:27.514 "data_offset": 2048, 00:16:27.514 "data_size": 63488 00:16:27.514 }, 00:16:27.514 { 00:16:27.514 "name": "BaseBdev3", 00:16:27.514 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:27.514 "is_configured": true, 00:16:27.514 "data_offset": 2048, 00:16:27.514 "data_size": 63488 00:16:27.514 }, 00:16:27.514 { 00:16:27.514 "name": "BaseBdev4", 00:16:27.514 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:27.514 "is_configured": true, 00:16:27.514 "data_offset": 2048, 00:16:27.514 "data_size": 63488 00:16:27.514 } 00:16:27.514 ] 00:16:27.514 }' 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:27.514 16:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:28.448 16:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:28.448 16:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:28.448 16:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:28.448 16:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:28.448 16:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:28.448 16:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:28.448 16:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:28.448 16:04:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:28.448 16:04:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:28.448 16:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:28.448 16:04:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:28.706 16:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:28.706 "name": "raid_bdev1", 00:16:28.706 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:28.706 "strip_size_kb": 64, 00:16:28.706 "state": "online", 00:16:28.706 "raid_level": "raid5f", 00:16:28.706 "superblock": true, 00:16:28.706 "num_base_bdevs": 4, 00:16:28.706 "num_base_bdevs_discovered": 4, 00:16:28.706 "num_base_bdevs_operational": 4, 00:16:28.706 "process": { 00:16:28.706 "type": "rebuild", 00:16:28.706 "target": "spare", 00:16:28.706 "progress": { 00:16:28.706 "blocks": 40320, 00:16:28.706 "percent": 21 00:16:28.706 } 00:16:28.706 }, 00:16:28.706 "base_bdevs_list": [ 00:16:28.706 { 00:16:28.706 "name": "spare", 00:16:28.706 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:28.706 "is_configured": true, 00:16:28.706 "data_offset": 2048, 00:16:28.706 "data_size": 63488 00:16:28.706 }, 00:16:28.706 { 00:16:28.706 "name": "BaseBdev2", 00:16:28.706 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:28.706 "is_configured": true, 00:16:28.706 "data_offset": 2048, 00:16:28.706 "data_size": 63488 00:16:28.706 }, 00:16:28.706 { 00:16:28.706 "name": "BaseBdev3", 00:16:28.706 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:28.706 "is_configured": true, 00:16:28.706 "data_offset": 2048, 00:16:28.706 "data_size": 63488 00:16:28.706 }, 00:16:28.706 { 00:16:28.706 "name": "BaseBdev4", 00:16:28.706 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:28.706 "is_configured": true, 00:16:28.706 "data_offset": 2048, 00:16:28.706 "data_size": 63488 00:16:28.706 } 00:16:28.706 ] 00:16:28.706 }' 00:16:28.706 16:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:28.706 16:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:28.706 16:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:28.706 16:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:28.706 16:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:29.638 16:04:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:29.638 16:04:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:29.638 16:04:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:29.638 16:04:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:29.638 16:04:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:29.638 16:04:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:29.638 16:04:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:29.638 16:04:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:29.638 16:04:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:29.638 16:04:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:29.638 16:04:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:29.638 16:04:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:29.638 "name": "raid_bdev1", 00:16:29.638 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:29.638 "strip_size_kb": 64, 00:16:29.638 "state": "online", 00:16:29.638 "raid_level": "raid5f", 00:16:29.638 "superblock": true, 00:16:29.638 "num_base_bdevs": 4, 00:16:29.638 "num_base_bdevs_discovered": 4, 00:16:29.638 "num_base_bdevs_operational": 4, 00:16:29.638 "process": { 00:16:29.638 "type": "rebuild", 00:16:29.638 "target": "spare", 00:16:29.638 "progress": { 00:16:29.638 "blocks": 61440, 00:16:29.638 "percent": 32 00:16:29.638 } 00:16:29.638 }, 00:16:29.638 "base_bdevs_list": [ 00:16:29.638 { 00:16:29.638 "name": "spare", 00:16:29.638 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:29.638 "is_configured": true, 00:16:29.638 "data_offset": 2048, 00:16:29.638 "data_size": 63488 00:16:29.638 }, 00:16:29.638 { 00:16:29.638 "name": "BaseBdev2", 00:16:29.638 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:29.638 "is_configured": true, 00:16:29.638 "data_offset": 2048, 00:16:29.638 "data_size": 63488 00:16:29.638 }, 00:16:29.638 { 00:16:29.638 "name": "BaseBdev3", 00:16:29.638 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:29.638 "is_configured": true, 00:16:29.638 "data_offset": 2048, 00:16:29.638 "data_size": 63488 00:16:29.638 }, 00:16:29.638 { 00:16:29.638 "name": "BaseBdev4", 00:16:29.638 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:29.638 "is_configured": true, 00:16:29.638 "data_offset": 2048, 00:16:29.638 "data_size": 63488 00:16:29.638 } 00:16:29.638 ] 00:16:29.638 }' 00:16:29.638 16:04:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:29.638 16:04:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:29.638 16:04:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:29.638 16:04:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:29.638 16:04:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:31.015 16:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:31.015 16:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:31.015 16:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:31.015 16:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:31.015 16:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:31.015 16:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:31.015 16:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:31.015 16:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:31.015 16:04:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:31.015 16:04:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:31.015 16:04:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.015 16:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:31.015 "name": "raid_bdev1", 00:16:31.015 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:31.015 "strip_size_kb": 64, 00:16:31.015 "state": "online", 00:16:31.015 "raid_level": "raid5f", 00:16:31.015 "superblock": true, 00:16:31.015 "num_base_bdevs": 4, 00:16:31.015 "num_base_bdevs_discovered": 4, 00:16:31.015 "num_base_bdevs_operational": 4, 00:16:31.016 "process": { 00:16:31.016 "type": "rebuild", 00:16:31.016 "target": "spare", 00:16:31.016 "progress": { 00:16:31.016 "blocks": 82560, 00:16:31.016 "percent": 43 00:16:31.016 } 00:16:31.016 }, 00:16:31.016 "base_bdevs_list": [ 00:16:31.016 { 00:16:31.016 "name": "spare", 00:16:31.016 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:31.016 "is_configured": true, 00:16:31.016 "data_offset": 2048, 00:16:31.016 "data_size": 63488 00:16:31.016 }, 00:16:31.016 { 00:16:31.016 "name": "BaseBdev2", 00:16:31.016 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:31.016 "is_configured": true, 00:16:31.016 "data_offset": 2048, 00:16:31.016 "data_size": 63488 00:16:31.016 }, 00:16:31.016 { 00:16:31.016 "name": "BaseBdev3", 00:16:31.016 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:31.016 "is_configured": true, 00:16:31.016 "data_offset": 2048, 00:16:31.016 "data_size": 63488 00:16:31.016 }, 00:16:31.016 { 00:16:31.016 "name": "BaseBdev4", 00:16:31.016 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:31.016 "is_configured": true, 00:16:31.016 "data_offset": 2048, 00:16:31.016 "data_size": 63488 00:16:31.016 } 00:16:31.016 ] 00:16:31.016 }' 00:16:31.016 16:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:31.016 16:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:31.016 16:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:31.016 16:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:31.016 16:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:31.947 16:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:31.947 16:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:31.947 16:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:31.947 16:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:31.947 16:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:31.947 16:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:31.948 16:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:31.948 16:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:31.948 16:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:31.948 16:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:31.948 16:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.948 16:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:31.948 "name": "raid_bdev1", 00:16:31.948 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:31.948 "strip_size_kb": 64, 00:16:31.948 "state": "online", 00:16:31.948 "raid_level": "raid5f", 00:16:31.948 "superblock": true, 00:16:31.948 "num_base_bdevs": 4, 00:16:31.948 "num_base_bdevs_discovered": 4, 00:16:31.948 "num_base_bdevs_operational": 4, 00:16:31.948 "process": { 00:16:31.948 "type": "rebuild", 00:16:31.948 "target": "spare", 00:16:31.948 "progress": { 00:16:31.948 "blocks": 103680, 00:16:31.948 "percent": 54 00:16:31.948 } 00:16:31.948 }, 00:16:31.948 "base_bdevs_list": [ 00:16:31.948 { 00:16:31.948 "name": "spare", 00:16:31.948 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:31.948 "is_configured": true, 00:16:31.948 "data_offset": 2048, 00:16:31.948 "data_size": 63488 00:16:31.948 }, 00:16:31.948 { 00:16:31.948 "name": "BaseBdev2", 00:16:31.948 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:31.948 "is_configured": true, 00:16:31.948 "data_offset": 2048, 00:16:31.948 "data_size": 63488 00:16:31.948 }, 00:16:31.948 { 00:16:31.948 "name": "BaseBdev3", 00:16:31.948 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:31.948 "is_configured": true, 00:16:31.948 "data_offset": 2048, 00:16:31.948 "data_size": 63488 00:16:31.948 }, 00:16:31.948 { 00:16:31.948 "name": "BaseBdev4", 00:16:31.948 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:31.948 "is_configured": true, 00:16:31.948 "data_offset": 2048, 00:16:31.948 "data_size": 63488 00:16:31.948 } 00:16:31.948 ] 00:16:31.948 }' 00:16:31.948 16:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:31.948 16:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:31.948 16:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:31.948 16:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:31.948 16:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:32.882 16:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:32.882 16:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:32.882 16:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:32.882 16:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:32.882 16:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:32.882 16:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:32.882 16:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:32.882 16:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:32.882 16:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:32.882 16:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:32.882 16:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:32.882 16:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:32.882 "name": "raid_bdev1", 00:16:32.882 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:32.882 "strip_size_kb": 64, 00:16:32.882 "state": "online", 00:16:32.882 "raid_level": "raid5f", 00:16:32.882 "superblock": true, 00:16:32.882 "num_base_bdevs": 4, 00:16:32.882 "num_base_bdevs_discovered": 4, 00:16:32.882 "num_base_bdevs_operational": 4, 00:16:32.882 "process": { 00:16:32.882 "type": "rebuild", 00:16:32.882 "target": "spare", 00:16:32.882 "progress": { 00:16:32.882 "blocks": 124800, 00:16:32.882 "percent": 65 00:16:32.882 } 00:16:32.882 }, 00:16:32.882 "base_bdevs_list": [ 00:16:32.882 { 00:16:32.882 "name": "spare", 00:16:32.882 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:32.882 "is_configured": true, 00:16:32.882 "data_offset": 2048, 00:16:32.882 "data_size": 63488 00:16:32.882 }, 00:16:32.882 { 00:16:32.882 "name": "BaseBdev2", 00:16:32.882 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:32.882 "is_configured": true, 00:16:32.882 "data_offset": 2048, 00:16:32.882 "data_size": 63488 00:16:32.882 }, 00:16:32.882 { 00:16:32.882 "name": "BaseBdev3", 00:16:32.882 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:32.882 "is_configured": true, 00:16:32.882 "data_offset": 2048, 00:16:32.882 "data_size": 63488 00:16:32.882 }, 00:16:32.882 { 00:16:32.882 "name": "BaseBdev4", 00:16:32.882 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:32.882 "is_configured": true, 00:16:32.882 "data_offset": 2048, 00:16:32.882 "data_size": 63488 00:16:32.882 } 00:16:32.882 ] 00:16:32.882 }' 00:16:32.882 16:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:33.188 16:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:33.188 16:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:33.188 16:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:33.188 16:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:34.153 16:04:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:34.153 16:04:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:34.153 16:04:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:34.153 16:04:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:34.153 16:04:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:34.154 16:04:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:34.154 16:04:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:34.154 16:04:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:34.154 16:04:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:34.154 16:04:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:34.154 16:04:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:34.154 16:04:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:34.154 "name": "raid_bdev1", 00:16:34.154 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:34.154 "strip_size_kb": 64, 00:16:34.154 "state": "online", 00:16:34.154 "raid_level": "raid5f", 00:16:34.154 "superblock": true, 00:16:34.154 "num_base_bdevs": 4, 00:16:34.154 "num_base_bdevs_discovered": 4, 00:16:34.154 "num_base_bdevs_operational": 4, 00:16:34.154 "process": { 00:16:34.154 "type": "rebuild", 00:16:34.154 "target": "spare", 00:16:34.154 "progress": { 00:16:34.154 "blocks": 145920, 00:16:34.154 "percent": 76 00:16:34.154 } 00:16:34.154 }, 00:16:34.154 "base_bdevs_list": [ 00:16:34.154 { 00:16:34.154 "name": "spare", 00:16:34.154 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:34.154 "is_configured": true, 00:16:34.154 "data_offset": 2048, 00:16:34.154 "data_size": 63488 00:16:34.154 }, 00:16:34.154 { 00:16:34.154 "name": "BaseBdev2", 00:16:34.154 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:34.154 "is_configured": true, 00:16:34.154 "data_offset": 2048, 00:16:34.154 "data_size": 63488 00:16:34.154 }, 00:16:34.154 { 00:16:34.154 "name": "BaseBdev3", 00:16:34.154 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:34.154 "is_configured": true, 00:16:34.154 "data_offset": 2048, 00:16:34.154 "data_size": 63488 00:16:34.154 }, 00:16:34.154 { 00:16:34.154 "name": "BaseBdev4", 00:16:34.154 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:34.154 "is_configured": true, 00:16:34.154 "data_offset": 2048, 00:16:34.154 "data_size": 63488 00:16:34.154 } 00:16:34.154 ] 00:16:34.154 }' 00:16:34.154 16:04:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:34.154 16:04:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:34.154 16:04:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:34.154 16:04:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:34.154 16:04:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:35.087 16:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:35.087 16:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:35.087 16:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:35.087 16:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:35.087 16:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:35.087 16:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:35.087 16:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:35.087 16:04:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:35.087 16:04:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:35.087 16:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:35.087 16:04:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:35.087 16:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:35.087 "name": "raid_bdev1", 00:16:35.087 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:35.087 "strip_size_kb": 64, 00:16:35.087 "state": "online", 00:16:35.087 "raid_level": "raid5f", 00:16:35.087 "superblock": true, 00:16:35.087 "num_base_bdevs": 4, 00:16:35.087 "num_base_bdevs_discovered": 4, 00:16:35.087 "num_base_bdevs_operational": 4, 00:16:35.087 "process": { 00:16:35.087 "type": "rebuild", 00:16:35.087 "target": "spare", 00:16:35.087 "progress": { 00:16:35.087 "blocks": 167040, 00:16:35.087 "percent": 87 00:16:35.087 } 00:16:35.087 }, 00:16:35.087 "base_bdevs_list": [ 00:16:35.087 { 00:16:35.087 "name": "spare", 00:16:35.087 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:35.087 "is_configured": true, 00:16:35.087 "data_offset": 2048, 00:16:35.087 "data_size": 63488 00:16:35.087 }, 00:16:35.087 { 00:16:35.087 "name": "BaseBdev2", 00:16:35.087 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:35.087 "is_configured": true, 00:16:35.087 "data_offset": 2048, 00:16:35.087 "data_size": 63488 00:16:35.087 }, 00:16:35.087 { 00:16:35.087 "name": "BaseBdev3", 00:16:35.087 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:35.087 "is_configured": true, 00:16:35.087 "data_offset": 2048, 00:16:35.087 "data_size": 63488 00:16:35.087 }, 00:16:35.087 { 00:16:35.087 "name": "BaseBdev4", 00:16:35.087 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:35.087 "is_configured": true, 00:16:35.087 "data_offset": 2048, 00:16:35.087 "data_size": 63488 00:16:35.087 } 00:16:35.087 ] 00:16:35.087 }' 00:16:35.087 16:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:35.346 16:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:35.346 16:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:35.346 16:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:35.346 16:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:36.280 16:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:36.280 16:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:36.280 16:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:36.280 16:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:36.280 16:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:36.280 16:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:36.280 16:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:36.280 16:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:36.280 16:04:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:36.280 16:04:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:36.280 16:04:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.280 16:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:36.280 "name": "raid_bdev1", 00:16:36.280 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:36.280 "strip_size_kb": 64, 00:16:36.280 "state": "online", 00:16:36.280 "raid_level": "raid5f", 00:16:36.280 "superblock": true, 00:16:36.280 "num_base_bdevs": 4, 00:16:36.280 "num_base_bdevs_discovered": 4, 00:16:36.280 "num_base_bdevs_operational": 4, 00:16:36.280 "process": { 00:16:36.280 "type": "rebuild", 00:16:36.280 "target": "spare", 00:16:36.280 "progress": { 00:16:36.280 "blocks": 188160, 00:16:36.280 "percent": 98 00:16:36.280 } 00:16:36.280 }, 00:16:36.280 "base_bdevs_list": [ 00:16:36.280 { 00:16:36.280 "name": "spare", 00:16:36.280 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:36.280 "is_configured": true, 00:16:36.280 "data_offset": 2048, 00:16:36.280 "data_size": 63488 00:16:36.280 }, 00:16:36.280 { 00:16:36.280 "name": "BaseBdev2", 00:16:36.280 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:36.280 "is_configured": true, 00:16:36.280 "data_offset": 2048, 00:16:36.280 "data_size": 63488 00:16:36.280 }, 00:16:36.280 { 00:16:36.280 "name": "BaseBdev3", 00:16:36.280 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:36.280 "is_configured": true, 00:16:36.280 "data_offset": 2048, 00:16:36.280 "data_size": 63488 00:16:36.280 }, 00:16:36.280 { 00:16:36.280 "name": "BaseBdev4", 00:16:36.280 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:36.280 "is_configured": true, 00:16:36.280 "data_offset": 2048, 00:16:36.280 "data_size": 63488 00:16:36.280 } 00:16:36.280 ] 00:16:36.280 }' 00:16:36.280 16:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:36.280 16:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:36.280 16:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:36.280 16:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:36.280 16:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:36.538 [2024-11-20 16:04:34.553639] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:36.538 [2024-11-20 16:04:34.553718] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:36.538 [2024-11-20 16:04:34.553837] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:37.472 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:37.472 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:37.472 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:37.472 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:37.472 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:37.472 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:37.472 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:37.472 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:37.472 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:37.472 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:37.472 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:37.473 "name": "raid_bdev1", 00:16:37.473 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:37.473 "strip_size_kb": 64, 00:16:37.473 "state": "online", 00:16:37.473 "raid_level": "raid5f", 00:16:37.473 "superblock": true, 00:16:37.473 "num_base_bdevs": 4, 00:16:37.473 "num_base_bdevs_discovered": 4, 00:16:37.473 "num_base_bdevs_operational": 4, 00:16:37.473 "base_bdevs_list": [ 00:16:37.473 { 00:16:37.473 "name": "spare", 00:16:37.473 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:37.473 "is_configured": true, 00:16:37.473 "data_offset": 2048, 00:16:37.473 "data_size": 63488 00:16:37.473 }, 00:16:37.473 { 00:16:37.473 "name": "BaseBdev2", 00:16:37.473 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:37.473 "is_configured": true, 00:16:37.473 "data_offset": 2048, 00:16:37.473 "data_size": 63488 00:16:37.473 }, 00:16:37.473 { 00:16:37.473 "name": "BaseBdev3", 00:16:37.473 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:37.473 "is_configured": true, 00:16:37.473 "data_offset": 2048, 00:16:37.473 "data_size": 63488 00:16:37.473 }, 00:16:37.473 { 00:16:37.473 "name": "BaseBdev4", 00:16:37.473 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:37.473 "is_configured": true, 00:16:37.473 "data_offset": 2048, 00:16:37.473 "data_size": 63488 00:16:37.473 } 00:16:37.473 ] 00:16:37.473 }' 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:37.473 "name": "raid_bdev1", 00:16:37.473 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:37.473 "strip_size_kb": 64, 00:16:37.473 "state": "online", 00:16:37.473 "raid_level": "raid5f", 00:16:37.473 "superblock": true, 00:16:37.473 "num_base_bdevs": 4, 00:16:37.473 "num_base_bdevs_discovered": 4, 00:16:37.473 "num_base_bdevs_operational": 4, 00:16:37.473 "base_bdevs_list": [ 00:16:37.473 { 00:16:37.473 "name": "spare", 00:16:37.473 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:37.473 "is_configured": true, 00:16:37.473 "data_offset": 2048, 00:16:37.473 "data_size": 63488 00:16:37.473 }, 00:16:37.473 { 00:16:37.473 "name": "BaseBdev2", 00:16:37.473 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:37.473 "is_configured": true, 00:16:37.473 "data_offset": 2048, 00:16:37.473 "data_size": 63488 00:16:37.473 }, 00:16:37.473 { 00:16:37.473 "name": "BaseBdev3", 00:16:37.473 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:37.473 "is_configured": true, 00:16:37.473 "data_offset": 2048, 00:16:37.473 "data_size": 63488 00:16:37.473 }, 00:16:37.473 { 00:16:37.473 "name": "BaseBdev4", 00:16:37.473 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:37.473 "is_configured": true, 00:16:37.473 "data_offset": 2048, 00:16:37.473 "data_size": 63488 00:16:37.473 } 00:16:37.473 ] 00:16:37.473 }' 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:37.473 "name": "raid_bdev1", 00:16:37.473 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:37.473 "strip_size_kb": 64, 00:16:37.473 "state": "online", 00:16:37.473 "raid_level": "raid5f", 00:16:37.473 "superblock": true, 00:16:37.473 "num_base_bdevs": 4, 00:16:37.473 "num_base_bdevs_discovered": 4, 00:16:37.473 "num_base_bdevs_operational": 4, 00:16:37.473 "base_bdevs_list": [ 00:16:37.473 { 00:16:37.473 "name": "spare", 00:16:37.473 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:37.473 "is_configured": true, 00:16:37.473 "data_offset": 2048, 00:16:37.473 "data_size": 63488 00:16:37.473 }, 00:16:37.473 { 00:16:37.473 "name": "BaseBdev2", 00:16:37.473 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:37.473 "is_configured": true, 00:16:37.473 "data_offset": 2048, 00:16:37.473 "data_size": 63488 00:16:37.473 }, 00:16:37.473 { 00:16:37.473 "name": "BaseBdev3", 00:16:37.473 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:37.473 "is_configured": true, 00:16:37.473 "data_offset": 2048, 00:16:37.473 "data_size": 63488 00:16:37.473 }, 00:16:37.473 { 00:16:37.473 "name": "BaseBdev4", 00:16:37.473 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:37.473 "is_configured": true, 00:16:37.473 "data_offset": 2048, 00:16:37.473 "data_size": 63488 00:16:37.473 } 00:16:37.473 ] 00:16:37.473 }' 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:37.473 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:37.731 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:37.731 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:37.731 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:37.731 [2024-11-20 16:04:35.970510] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:37.731 [2024-11-20 16:04:35.970535] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:37.731 [2024-11-20 16:04:35.970604] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:37.731 [2024-11-20 16:04:35.970695] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:37.731 [2024-11-20 16:04:35.970706] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:16:37.731 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:37.731 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:37.731 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:16:37.731 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:37.731 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:37.988 16:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:16:37.988 /dev/nbd0 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:37.988 1+0 records in 00:16:37.988 1+0 records out 00:16:37.988 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00019082 s, 21.5 MB/s 00:16:37.988 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:38.246 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:16:38.246 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:38.246 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:38.246 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:16:38.246 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:38.246 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:16:38.247 /dev/nbd1 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:38.247 1+0 records in 00:16:38.247 1+0 records out 00:16:38.247 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000228081 s, 18.0 MB/s 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:38.247 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:16:38.504 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:16:38.505 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:38.505 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:38.505 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:38.505 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:16:38.505 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:38.505 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:38.762 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:38.762 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:38.762 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:38.762 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:38.762 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:38.762 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:38.762 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:16:38.762 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:16:38.762 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:38.762 16:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.020 [2024-11-20 16:04:37.046580] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:39.020 [2024-11-20 16:04:37.046629] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:39.020 [2024-11-20 16:04:37.046651] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:16:39.020 [2024-11-20 16:04:37.046660] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:39.020 [2024-11-20 16:04:37.048926] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:39.020 [2024-11-20 16:04:37.049074] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:39.020 [2024-11-20 16:04:37.049178] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:39.020 [2024-11-20 16:04:37.049227] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:39.020 [2024-11-20 16:04:37.049360] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:39.020 [2024-11-20 16:04:37.049447] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:16:39.020 [2024-11-20 16:04:37.049525] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:16:39.020 spare 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.020 [2024-11-20 16:04:37.149616] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:16:39.020 [2024-11-20 16:04:37.149661] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:16:39.020 [2024-11-20 16:04:37.149979] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000491d0 00:16:39.020 [2024-11-20 16:04:37.154506] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:16:39.020 [2024-11-20 16:04:37.154525] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:16:39.020 [2024-11-20 16:04:37.154727] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:39.020 "name": "raid_bdev1", 00:16:39.020 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:39.020 "strip_size_kb": 64, 00:16:39.020 "state": "online", 00:16:39.020 "raid_level": "raid5f", 00:16:39.020 "superblock": true, 00:16:39.020 "num_base_bdevs": 4, 00:16:39.020 "num_base_bdevs_discovered": 4, 00:16:39.020 "num_base_bdevs_operational": 4, 00:16:39.020 "base_bdevs_list": [ 00:16:39.020 { 00:16:39.020 "name": "spare", 00:16:39.020 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:39.020 "is_configured": true, 00:16:39.020 "data_offset": 2048, 00:16:39.020 "data_size": 63488 00:16:39.020 }, 00:16:39.020 { 00:16:39.020 "name": "BaseBdev2", 00:16:39.020 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:39.020 "is_configured": true, 00:16:39.020 "data_offset": 2048, 00:16:39.020 "data_size": 63488 00:16:39.020 }, 00:16:39.020 { 00:16:39.020 "name": "BaseBdev3", 00:16:39.020 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:39.020 "is_configured": true, 00:16:39.020 "data_offset": 2048, 00:16:39.020 "data_size": 63488 00:16:39.020 }, 00:16:39.020 { 00:16:39.020 "name": "BaseBdev4", 00:16:39.020 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:39.020 "is_configured": true, 00:16:39.020 "data_offset": 2048, 00:16:39.020 "data_size": 63488 00:16:39.020 } 00:16:39.020 ] 00:16:39.020 }' 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:39.020 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.278 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:39.278 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:39.278 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:39.278 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:39.278 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:39.278 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:39.278 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:39.278 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.278 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.278 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.278 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:39.278 "name": "raid_bdev1", 00:16:39.278 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:39.278 "strip_size_kb": 64, 00:16:39.278 "state": "online", 00:16:39.278 "raid_level": "raid5f", 00:16:39.278 "superblock": true, 00:16:39.278 "num_base_bdevs": 4, 00:16:39.278 "num_base_bdevs_discovered": 4, 00:16:39.278 "num_base_bdevs_operational": 4, 00:16:39.278 "base_bdevs_list": [ 00:16:39.278 { 00:16:39.278 "name": "spare", 00:16:39.278 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:39.278 "is_configured": true, 00:16:39.278 "data_offset": 2048, 00:16:39.278 "data_size": 63488 00:16:39.278 }, 00:16:39.278 { 00:16:39.278 "name": "BaseBdev2", 00:16:39.278 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:39.278 "is_configured": true, 00:16:39.278 "data_offset": 2048, 00:16:39.278 "data_size": 63488 00:16:39.278 }, 00:16:39.278 { 00:16:39.278 "name": "BaseBdev3", 00:16:39.278 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:39.278 "is_configured": true, 00:16:39.278 "data_offset": 2048, 00:16:39.278 "data_size": 63488 00:16:39.278 }, 00:16:39.278 { 00:16:39.278 "name": "BaseBdev4", 00:16:39.278 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:39.278 "is_configured": true, 00:16:39.278 "data_offset": 2048, 00:16:39.278 "data_size": 63488 00:16:39.278 } 00:16:39.278 ] 00:16:39.278 }' 00:16:39.278 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:39.537 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:39.537 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:39.537 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:39.537 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:39.537 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.537 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.537 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:16:39.537 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.537 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:16:39.537 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:39.537 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.537 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.538 [2024-11-20 16:04:37.608083] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:39.538 "name": "raid_bdev1", 00:16:39.538 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:39.538 "strip_size_kb": 64, 00:16:39.538 "state": "online", 00:16:39.538 "raid_level": "raid5f", 00:16:39.538 "superblock": true, 00:16:39.538 "num_base_bdevs": 4, 00:16:39.538 "num_base_bdevs_discovered": 3, 00:16:39.538 "num_base_bdevs_operational": 3, 00:16:39.538 "base_bdevs_list": [ 00:16:39.538 { 00:16:39.538 "name": null, 00:16:39.538 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:39.538 "is_configured": false, 00:16:39.538 "data_offset": 0, 00:16:39.538 "data_size": 63488 00:16:39.538 }, 00:16:39.538 { 00:16:39.538 "name": "BaseBdev2", 00:16:39.538 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:39.538 "is_configured": true, 00:16:39.538 "data_offset": 2048, 00:16:39.538 "data_size": 63488 00:16:39.538 }, 00:16:39.538 { 00:16:39.538 "name": "BaseBdev3", 00:16:39.538 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:39.538 "is_configured": true, 00:16:39.538 "data_offset": 2048, 00:16:39.538 "data_size": 63488 00:16:39.538 }, 00:16:39.538 { 00:16:39.538 "name": "BaseBdev4", 00:16:39.538 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:39.538 "is_configured": true, 00:16:39.538 "data_offset": 2048, 00:16:39.538 "data_size": 63488 00:16:39.538 } 00:16:39.538 ] 00:16:39.538 }' 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:39.538 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.801 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:39.801 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.801 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.801 [2024-11-20 16:04:37.940242] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:39.801 [2024-11-20 16:04:37.940539] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:39.801 [2024-11-20 16:04:37.940663] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:39.801 [2024-11-20 16:04:37.940777] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:39.801 [2024-11-20 16:04:37.950894] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000492a0 00:16:39.801 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.801 16:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:16:39.801 [2024-11-20 16:04:37.958115] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:40.734 16:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:40.734 16:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:40.734 16:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:40.734 16:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:40.734 16:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:40.734 16:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:40.734 16:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:40.734 16:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:40.734 16:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:40.734 16:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:40.734 16:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:40.734 "name": "raid_bdev1", 00:16:40.734 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:40.734 "strip_size_kb": 64, 00:16:40.734 "state": "online", 00:16:40.734 "raid_level": "raid5f", 00:16:40.734 "superblock": true, 00:16:40.734 "num_base_bdevs": 4, 00:16:40.734 "num_base_bdevs_discovered": 4, 00:16:40.734 "num_base_bdevs_operational": 4, 00:16:40.734 "process": { 00:16:40.734 "type": "rebuild", 00:16:40.734 "target": "spare", 00:16:40.734 "progress": { 00:16:40.734 "blocks": 17280, 00:16:40.734 "percent": 9 00:16:40.734 } 00:16:40.734 }, 00:16:40.734 "base_bdevs_list": [ 00:16:40.734 { 00:16:40.734 "name": "spare", 00:16:40.734 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:40.734 "is_configured": true, 00:16:40.734 "data_offset": 2048, 00:16:40.734 "data_size": 63488 00:16:40.734 }, 00:16:40.734 { 00:16:40.734 "name": "BaseBdev2", 00:16:40.734 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:40.734 "is_configured": true, 00:16:40.734 "data_offset": 2048, 00:16:40.734 "data_size": 63488 00:16:40.734 }, 00:16:40.734 { 00:16:40.734 "name": "BaseBdev3", 00:16:40.734 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:40.734 "is_configured": true, 00:16:40.734 "data_offset": 2048, 00:16:40.734 "data_size": 63488 00:16:40.734 }, 00:16:40.734 { 00:16:40.734 "name": "BaseBdev4", 00:16:40.734 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:40.734 "is_configured": true, 00:16:40.734 "data_offset": 2048, 00:16:40.734 "data_size": 63488 00:16:40.734 } 00:16:40.734 ] 00:16:40.734 }' 00:16:40.992 16:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:40.992 [2024-11-20 16:04:39.043090] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:40.992 [2024-11-20 16:04:39.066789] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:40.992 [2024-11-20 16:04:39.066965] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:40.992 [2024-11-20 16:04:39.067105] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:40.992 [2024-11-20 16:04:39.067134] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:40.992 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:40.992 "name": "raid_bdev1", 00:16:40.992 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:40.992 "strip_size_kb": 64, 00:16:40.992 "state": "online", 00:16:40.992 "raid_level": "raid5f", 00:16:40.992 "superblock": true, 00:16:40.992 "num_base_bdevs": 4, 00:16:40.992 "num_base_bdevs_discovered": 3, 00:16:40.992 "num_base_bdevs_operational": 3, 00:16:40.992 "base_bdevs_list": [ 00:16:40.992 { 00:16:40.992 "name": null, 00:16:40.992 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:40.992 "is_configured": false, 00:16:40.992 "data_offset": 0, 00:16:40.992 "data_size": 63488 00:16:40.992 }, 00:16:40.992 { 00:16:40.992 "name": "BaseBdev2", 00:16:40.992 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:40.992 "is_configured": true, 00:16:40.992 "data_offset": 2048, 00:16:40.992 "data_size": 63488 00:16:40.992 }, 00:16:40.992 { 00:16:40.992 "name": "BaseBdev3", 00:16:40.992 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:40.992 "is_configured": true, 00:16:40.992 "data_offset": 2048, 00:16:40.992 "data_size": 63488 00:16:40.992 }, 00:16:40.992 { 00:16:40.992 "name": "BaseBdev4", 00:16:40.992 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:40.992 "is_configured": true, 00:16:40.992 "data_offset": 2048, 00:16:40.992 "data_size": 63488 00:16:40.992 } 00:16:40.992 ] 00:16:40.992 }' 00:16:40.993 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:40.993 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:41.249 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:41.249 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:41.249 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:41.249 [2024-11-20 16:04:39.406244] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:41.249 [2024-11-20 16:04:39.406417] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:41.249 [2024-11-20 16:04:39.406505] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c380 00:16:41.249 [2024-11-20 16:04:39.406566] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:41.250 [2024-11-20 16:04:39.407056] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:41.250 [2024-11-20 16:04:39.407084] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:41.250 [2024-11-20 16:04:39.407170] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:41.250 [2024-11-20 16:04:39.407185] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:41.250 [2024-11-20 16:04:39.407194] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:41.250 [2024-11-20 16:04:39.407221] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:41.250 [2024-11-20 16:04:39.416561] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000049370 00:16:41.250 spare 00:16:41.250 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:41.250 16:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:16:41.250 [2024-11-20 16:04:39.422814] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:42.183 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:42.183 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:42.183 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:42.183 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:42.183 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:42.183 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:42.183 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:42.183 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:42.183 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:42.441 "name": "raid_bdev1", 00:16:42.441 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:42.441 "strip_size_kb": 64, 00:16:42.441 "state": "online", 00:16:42.441 "raid_level": "raid5f", 00:16:42.441 "superblock": true, 00:16:42.441 "num_base_bdevs": 4, 00:16:42.441 "num_base_bdevs_discovered": 4, 00:16:42.441 "num_base_bdevs_operational": 4, 00:16:42.441 "process": { 00:16:42.441 "type": "rebuild", 00:16:42.441 "target": "spare", 00:16:42.441 "progress": { 00:16:42.441 "blocks": 19200, 00:16:42.441 "percent": 10 00:16:42.441 } 00:16:42.441 }, 00:16:42.441 "base_bdevs_list": [ 00:16:42.441 { 00:16:42.441 "name": "spare", 00:16:42.441 "uuid": "56e80fcb-5735-5ed6-a5e0-27d3a3feb80c", 00:16:42.441 "is_configured": true, 00:16:42.441 "data_offset": 2048, 00:16:42.441 "data_size": 63488 00:16:42.441 }, 00:16:42.441 { 00:16:42.441 "name": "BaseBdev2", 00:16:42.441 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:42.441 "is_configured": true, 00:16:42.441 "data_offset": 2048, 00:16:42.441 "data_size": 63488 00:16:42.441 }, 00:16:42.441 { 00:16:42.441 "name": "BaseBdev3", 00:16:42.441 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:42.441 "is_configured": true, 00:16:42.441 "data_offset": 2048, 00:16:42.441 "data_size": 63488 00:16:42.441 }, 00:16:42.441 { 00:16:42.441 "name": "BaseBdev4", 00:16:42.441 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:42.441 "is_configured": true, 00:16:42.441 "data_offset": 2048, 00:16:42.441 "data_size": 63488 00:16:42.441 } 00:16:42.441 ] 00:16:42.441 }' 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:42.441 [2024-11-20 16:04:40.519752] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:42.441 [2024-11-20 16:04:40.531339] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:42.441 [2024-11-20 16:04:40.531393] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:42.441 [2024-11-20 16:04:40.531412] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:42.441 [2024-11-20 16:04:40.531419] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:42.441 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:42.441 "name": "raid_bdev1", 00:16:42.441 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:42.441 "strip_size_kb": 64, 00:16:42.441 "state": "online", 00:16:42.441 "raid_level": "raid5f", 00:16:42.441 "superblock": true, 00:16:42.441 "num_base_bdevs": 4, 00:16:42.441 "num_base_bdevs_discovered": 3, 00:16:42.441 "num_base_bdevs_operational": 3, 00:16:42.441 "base_bdevs_list": [ 00:16:42.441 { 00:16:42.441 "name": null, 00:16:42.441 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:42.441 "is_configured": false, 00:16:42.441 "data_offset": 0, 00:16:42.441 "data_size": 63488 00:16:42.441 }, 00:16:42.441 { 00:16:42.441 "name": "BaseBdev2", 00:16:42.441 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:42.441 "is_configured": true, 00:16:42.441 "data_offset": 2048, 00:16:42.441 "data_size": 63488 00:16:42.441 }, 00:16:42.441 { 00:16:42.441 "name": "BaseBdev3", 00:16:42.441 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:42.442 "is_configured": true, 00:16:42.442 "data_offset": 2048, 00:16:42.442 "data_size": 63488 00:16:42.442 }, 00:16:42.442 { 00:16:42.442 "name": "BaseBdev4", 00:16:42.442 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:42.442 "is_configured": true, 00:16:42.442 "data_offset": 2048, 00:16:42.442 "data_size": 63488 00:16:42.442 } 00:16:42.442 ] 00:16:42.442 }' 00:16:42.442 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:42.442 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:42.699 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:42.699 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:42.699 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:42.699 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:42.699 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:42.699 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:42.699 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:42.699 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:42.699 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:42.699 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:42.699 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:42.699 "name": "raid_bdev1", 00:16:42.699 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:42.699 "strip_size_kb": 64, 00:16:42.699 "state": "online", 00:16:42.699 "raid_level": "raid5f", 00:16:42.700 "superblock": true, 00:16:42.700 "num_base_bdevs": 4, 00:16:42.700 "num_base_bdevs_discovered": 3, 00:16:42.700 "num_base_bdevs_operational": 3, 00:16:42.700 "base_bdevs_list": [ 00:16:42.700 { 00:16:42.700 "name": null, 00:16:42.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:42.700 "is_configured": false, 00:16:42.700 "data_offset": 0, 00:16:42.700 "data_size": 63488 00:16:42.700 }, 00:16:42.700 { 00:16:42.700 "name": "BaseBdev2", 00:16:42.700 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:42.700 "is_configured": true, 00:16:42.700 "data_offset": 2048, 00:16:42.700 "data_size": 63488 00:16:42.700 }, 00:16:42.700 { 00:16:42.700 "name": "BaseBdev3", 00:16:42.700 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:42.700 "is_configured": true, 00:16:42.700 "data_offset": 2048, 00:16:42.700 "data_size": 63488 00:16:42.700 }, 00:16:42.700 { 00:16:42.700 "name": "BaseBdev4", 00:16:42.700 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:42.700 "is_configured": true, 00:16:42.700 "data_offset": 2048, 00:16:42.700 "data_size": 63488 00:16:42.700 } 00:16:42.700 ] 00:16:42.700 }' 00:16:42.700 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:42.958 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:42.958 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:42.958 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:42.958 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:16:42.958 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:42.958 16:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:42.958 16:04:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:42.958 16:04:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:42.958 16:04:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:42.958 16:04:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:42.958 [2024-11-20 16:04:41.006474] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:42.958 [2024-11-20 16:04:41.006677] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:42.958 [2024-11-20 16:04:41.006705] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c980 00:16:42.958 [2024-11-20 16:04:41.006714] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:42.958 [2024-11-20 16:04:41.007139] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:42.958 [2024-11-20 16:04:41.007163] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:42.958 [2024-11-20 16:04:41.007232] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:16:42.958 [2024-11-20 16:04:41.007249] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:42.958 [2024-11-20 16:04:41.007260] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:42.958 [2024-11-20 16:04:41.007269] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:16:42.958 BaseBdev1 00:16:42.958 16:04:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:42.958 16:04:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:43.969 "name": "raid_bdev1", 00:16:43.969 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:43.969 "strip_size_kb": 64, 00:16:43.969 "state": "online", 00:16:43.969 "raid_level": "raid5f", 00:16:43.969 "superblock": true, 00:16:43.969 "num_base_bdevs": 4, 00:16:43.969 "num_base_bdevs_discovered": 3, 00:16:43.969 "num_base_bdevs_operational": 3, 00:16:43.969 "base_bdevs_list": [ 00:16:43.969 { 00:16:43.969 "name": null, 00:16:43.969 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:43.969 "is_configured": false, 00:16:43.969 "data_offset": 0, 00:16:43.969 "data_size": 63488 00:16:43.969 }, 00:16:43.969 { 00:16:43.969 "name": "BaseBdev2", 00:16:43.969 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:43.969 "is_configured": true, 00:16:43.969 "data_offset": 2048, 00:16:43.969 "data_size": 63488 00:16:43.969 }, 00:16:43.969 { 00:16:43.969 "name": "BaseBdev3", 00:16:43.969 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:43.969 "is_configured": true, 00:16:43.969 "data_offset": 2048, 00:16:43.969 "data_size": 63488 00:16:43.969 }, 00:16:43.969 { 00:16:43.969 "name": "BaseBdev4", 00:16:43.969 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:43.969 "is_configured": true, 00:16:43.969 "data_offset": 2048, 00:16:43.969 "data_size": 63488 00:16:43.969 } 00:16:43.969 ] 00:16:43.969 }' 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:43.969 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:44.228 "name": "raid_bdev1", 00:16:44.228 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:44.228 "strip_size_kb": 64, 00:16:44.228 "state": "online", 00:16:44.228 "raid_level": "raid5f", 00:16:44.228 "superblock": true, 00:16:44.228 "num_base_bdevs": 4, 00:16:44.228 "num_base_bdevs_discovered": 3, 00:16:44.228 "num_base_bdevs_operational": 3, 00:16:44.228 "base_bdevs_list": [ 00:16:44.228 { 00:16:44.228 "name": null, 00:16:44.228 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:44.228 "is_configured": false, 00:16:44.228 "data_offset": 0, 00:16:44.228 "data_size": 63488 00:16:44.228 }, 00:16:44.228 { 00:16:44.228 "name": "BaseBdev2", 00:16:44.228 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:44.228 "is_configured": true, 00:16:44.228 "data_offset": 2048, 00:16:44.228 "data_size": 63488 00:16:44.228 }, 00:16:44.228 { 00:16:44.228 "name": "BaseBdev3", 00:16:44.228 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:44.228 "is_configured": true, 00:16:44.228 "data_offset": 2048, 00:16:44.228 "data_size": 63488 00:16:44.228 }, 00:16:44.228 { 00:16:44.228 "name": "BaseBdev4", 00:16:44.228 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:44.228 "is_configured": true, 00:16:44.228 "data_offset": 2048, 00:16:44.228 "data_size": 63488 00:16:44.228 } 00:16:44.228 ] 00:16:44.228 }' 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:44.228 [2024-11-20 16:04:42.406851] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:44.228 [2024-11-20 16:04:42.406981] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:44.228 [2024-11-20 16:04:42.406993] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:44.228 request: 00:16:44.228 { 00:16:44.228 "base_bdev": "BaseBdev1", 00:16:44.228 "raid_bdev": "raid_bdev1", 00:16:44.228 "method": "bdev_raid_add_base_bdev", 00:16:44.228 "req_id": 1 00:16:44.228 } 00:16:44.228 Got JSON-RPC error response 00:16:44.228 response: 00:16:44.228 { 00:16:44.228 "code": -22, 00:16:44.228 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:16:44.228 } 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:16:44.228 16:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:45.601 "name": "raid_bdev1", 00:16:45.601 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:45.601 "strip_size_kb": 64, 00:16:45.601 "state": "online", 00:16:45.601 "raid_level": "raid5f", 00:16:45.601 "superblock": true, 00:16:45.601 "num_base_bdevs": 4, 00:16:45.601 "num_base_bdevs_discovered": 3, 00:16:45.601 "num_base_bdevs_operational": 3, 00:16:45.601 "base_bdevs_list": [ 00:16:45.601 { 00:16:45.601 "name": null, 00:16:45.601 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:45.601 "is_configured": false, 00:16:45.601 "data_offset": 0, 00:16:45.601 "data_size": 63488 00:16:45.601 }, 00:16:45.601 { 00:16:45.601 "name": "BaseBdev2", 00:16:45.601 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:45.601 "is_configured": true, 00:16:45.601 "data_offset": 2048, 00:16:45.601 "data_size": 63488 00:16:45.601 }, 00:16:45.601 { 00:16:45.601 "name": "BaseBdev3", 00:16:45.601 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:45.601 "is_configured": true, 00:16:45.601 "data_offset": 2048, 00:16:45.601 "data_size": 63488 00:16:45.601 }, 00:16:45.601 { 00:16:45.601 "name": "BaseBdev4", 00:16:45.601 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:45.601 "is_configured": true, 00:16:45.601 "data_offset": 2048, 00:16:45.601 "data_size": 63488 00:16:45.601 } 00:16:45.601 ] 00:16:45.601 }' 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:45.601 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:45.601 "name": "raid_bdev1", 00:16:45.601 "uuid": "fa82a5b5-d900-47e7-94b6-3b05a7ca4408", 00:16:45.601 "strip_size_kb": 64, 00:16:45.601 "state": "online", 00:16:45.601 "raid_level": "raid5f", 00:16:45.601 "superblock": true, 00:16:45.601 "num_base_bdevs": 4, 00:16:45.601 "num_base_bdevs_discovered": 3, 00:16:45.601 "num_base_bdevs_operational": 3, 00:16:45.601 "base_bdevs_list": [ 00:16:45.601 { 00:16:45.601 "name": null, 00:16:45.601 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:45.601 "is_configured": false, 00:16:45.601 "data_offset": 0, 00:16:45.601 "data_size": 63488 00:16:45.601 }, 00:16:45.601 { 00:16:45.601 "name": "BaseBdev2", 00:16:45.602 "uuid": "3035552f-b2d9-5681-b39c-824be11fea17", 00:16:45.602 "is_configured": true, 00:16:45.602 "data_offset": 2048, 00:16:45.602 "data_size": 63488 00:16:45.602 }, 00:16:45.602 { 00:16:45.602 "name": "BaseBdev3", 00:16:45.602 "uuid": "3ed4ed6f-18c5-5924-a5f3-7da299598b58", 00:16:45.602 "is_configured": true, 00:16:45.602 "data_offset": 2048, 00:16:45.602 "data_size": 63488 00:16:45.602 }, 00:16:45.602 { 00:16:45.602 "name": "BaseBdev4", 00:16:45.602 "uuid": "b472aa53-30e4-55ce-9f22-9c6ec77cc96b", 00:16:45.602 "is_configured": true, 00:16:45.602 "data_offset": 2048, 00:16:45.602 "data_size": 63488 00:16:45.602 } 00:16:45.602 ] 00:16:45.602 }' 00:16:45.602 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:45.602 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:45.602 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:45.602 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:45.602 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 82724 00:16:45.602 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 82724 ']' 00:16:45.602 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 82724 00:16:45.602 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:16:45.602 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:45.602 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 82724 00:16:45.602 killing process with pid 82724 00:16:45.602 Received shutdown signal, test time was about 60.000000 seconds 00:16:45.602 00:16:45.602 Latency(us) 00:16:45.602 [2024-11-20T16:04:43.853Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:45.602 [2024-11-20T16:04:43.853Z] =================================================================================================================== 00:16:45.602 [2024-11-20T16:04:43.853Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:45.602 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:45.602 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:45.602 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 82724' 00:16:45.602 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 82724 00:16:45.602 [2024-11-20 16:04:43.831059] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:45.602 16:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 82724 00:16:45.602 [2024-11-20 16:04:43.831152] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:45.602 [2024-11-20 16:04:43.831214] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:45.602 [2024-11-20 16:04:43.831226] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:16:45.858 [2024-11-20 16:04:44.069203] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:46.424 16:04:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:16:46.424 00:16:46.424 real 0m24.437s 00:16:46.424 user 0m29.626s 00:16:46.424 sys 0m2.131s 00:16:46.424 16:04:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:46.424 ************************************ 00:16:46.424 END TEST raid5f_rebuild_test_sb 00:16:46.424 ************************************ 00:16:46.424 16:04:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:46.424 16:04:44 bdev_raid -- bdev/bdev_raid.sh@995 -- # base_blocklen=4096 00:16:46.424 16:04:44 bdev_raid -- bdev/bdev_raid.sh@997 -- # run_test raid_state_function_test_sb_4k raid_state_function_test raid1 2 true 00:16:46.424 16:04:44 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:16:46.424 16:04:44 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:46.424 16:04:44 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:46.682 ************************************ 00:16:46.682 START TEST raid_state_function_test_sb_4k 00:16:46.682 ************************************ 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # local strip_size 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:16:46.682 Process raid pid: 83525 00:16:46.682 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@229 -- # raid_pid=83525 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 83525' 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@231 -- # waitforlisten 83525 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@835 -- # '[' -z 83525 ']' 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:46.682 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:16:46.683 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:46.683 16:04:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:46.683 [2024-11-20 16:04:44.747409] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:16:46.683 [2024-11-20 16:04:44.747690] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:46.683 [2024-11-20 16:04:44.904026] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:46.939 [2024-11-20 16:04:44.990272] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:46.939 [2024-11-20 16:04:45.102126] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:46.939 [2024-11-20 16:04:45.102160] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@868 -- # return 0 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:47.543 [2024-11-20 16:04:45.602360] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:47.543 [2024-11-20 16:04:45.602407] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:47.543 [2024-11-20 16:04:45.602415] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:47.543 [2024-11-20 16:04:45.602423] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:47.543 "name": "Existed_Raid", 00:16:47.543 "uuid": "efad7604-a783-4f92-a295-b6e6604dcc44", 00:16:47.543 "strip_size_kb": 0, 00:16:47.543 "state": "configuring", 00:16:47.543 "raid_level": "raid1", 00:16:47.543 "superblock": true, 00:16:47.543 "num_base_bdevs": 2, 00:16:47.543 "num_base_bdevs_discovered": 0, 00:16:47.543 "num_base_bdevs_operational": 2, 00:16:47.543 "base_bdevs_list": [ 00:16:47.543 { 00:16:47.543 "name": "BaseBdev1", 00:16:47.543 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:47.543 "is_configured": false, 00:16:47.543 "data_offset": 0, 00:16:47.543 "data_size": 0 00:16:47.543 }, 00:16:47.543 { 00:16:47.543 "name": "BaseBdev2", 00:16:47.543 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:47.543 "is_configured": false, 00:16:47.543 "data_offset": 0, 00:16:47.543 "data_size": 0 00:16:47.543 } 00:16:47.543 ] 00:16:47.543 }' 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:47.543 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:47.802 [2024-11-20 16:04:45.922362] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:47.802 [2024-11-20 16:04:45.922388] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:47.802 [2024-11-20 16:04:45.930361] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:47.802 [2024-11-20 16:04:45.930393] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:47.802 [2024-11-20 16:04:45.930401] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:47.802 [2024-11-20 16:04:45.930410] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:47.802 [2024-11-20 16:04:45.958566] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:47.802 BaseBdev1 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@905 -- # local i 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:47.802 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:47.802 [ 00:16:47.802 { 00:16:47.802 "name": "BaseBdev1", 00:16:47.802 "aliases": [ 00:16:47.802 "d3562751-10c9-4e82-8d2c-66a0afe5d883" 00:16:47.802 ], 00:16:47.802 "product_name": "Malloc disk", 00:16:47.802 "block_size": 4096, 00:16:47.802 "num_blocks": 8192, 00:16:47.802 "uuid": "d3562751-10c9-4e82-8d2c-66a0afe5d883", 00:16:47.802 "assigned_rate_limits": { 00:16:47.802 "rw_ios_per_sec": 0, 00:16:47.802 "rw_mbytes_per_sec": 0, 00:16:47.802 "r_mbytes_per_sec": 0, 00:16:47.802 "w_mbytes_per_sec": 0 00:16:47.802 }, 00:16:47.802 "claimed": true, 00:16:47.802 "claim_type": "exclusive_write", 00:16:47.802 "zoned": false, 00:16:47.802 "supported_io_types": { 00:16:47.802 "read": true, 00:16:47.802 "write": true, 00:16:47.802 "unmap": true, 00:16:47.802 "flush": true, 00:16:47.802 "reset": true, 00:16:47.802 "nvme_admin": false, 00:16:47.802 "nvme_io": false, 00:16:47.802 "nvme_io_md": false, 00:16:47.802 "write_zeroes": true, 00:16:47.802 "zcopy": true, 00:16:47.802 "get_zone_info": false, 00:16:47.802 "zone_management": false, 00:16:47.802 "zone_append": false, 00:16:47.802 "compare": false, 00:16:47.802 "compare_and_write": false, 00:16:47.802 "abort": true, 00:16:47.802 "seek_hole": false, 00:16:47.802 "seek_data": false, 00:16:47.802 "copy": true, 00:16:47.802 "nvme_iov_md": false 00:16:47.802 }, 00:16:47.802 "memory_domains": [ 00:16:47.802 { 00:16:47.802 "dma_device_id": "system", 00:16:47.802 "dma_device_type": 1 00:16:47.802 }, 00:16:47.802 { 00:16:47.802 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:47.802 "dma_device_type": 2 00:16:47.802 } 00:16:47.802 ], 00:16:47.803 "driver_specific": {} 00:16:47.803 } 00:16:47.803 ] 00:16:47.803 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:47.803 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@911 -- # return 0 00:16:47.803 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:47.803 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:47.803 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:47.803 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:47.803 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:47.803 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:47.803 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:47.803 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:47.803 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:47.803 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:47.803 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:47.803 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:47.803 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:47.803 16:04:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:47.803 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:47.803 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:47.803 "name": "Existed_Raid", 00:16:47.803 "uuid": "0915d181-ec16-4dec-a55f-294a42b8d11e", 00:16:47.803 "strip_size_kb": 0, 00:16:47.803 "state": "configuring", 00:16:47.803 "raid_level": "raid1", 00:16:47.803 "superblock": true, 00:16:47.803 "num_base_bdevs": 2, 00:16:47.803 "num_base_bdevs_discovered": 1, 00:16:47.803 "num_base_bdevs_operational": 2, 00:16:47.803 "base_bdevs_list": [ 00:16:47.803 { 00:16:47.803 "name": "BaseBdev1", 00:16:47.803 "uuid": "d3562751-10c9-4e82-8d2c-66a0afe5d883", 00:16:47.803 "is_configured": true, 00:16:47.803 "data_offset": 256, 00:16:47.803 "data_size": 7936 00:16:47.803 }, 00:16:47.803 { 00:16:47.803 "name": "BaseBdev2", 00:16:47.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:47.803 "is_configured": false, 00:16:47.803 "data_offset": 0, 00:16:47.803 "data_size": 0 00:16:47.803 } 00:16:47.803 ] 00:16:47.803 }' 00:16:47.803 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:47.803 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.061 [2024-11-20 16:04:46.294677] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:48.061 [2024-11-20 16:04:46.294714] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.061 [2024-11-20 16:04:46.302717] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:48.061 [2024-11-20 16:04:46.304276] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:48.061 [2024-11-20 16:04:46.304312] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:48.061 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:48.319 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:48.319 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.319 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:48.319 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.319 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.319 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:48.319 "name": "Existed_Raid", 00:16:48.319 "uuid": "b2623de6-2f65-4d81-9245-8f1c3734e778", 00:16:48.319 "strip_size_kb": 0, 00:16:48.319 "state": "configuring", 00:16:48.319 "raid_level": "raid1", 00:16:48.319 "superblock": true, 00:16:48.319 "num_base_bdevs": 2, 00:16:48.319 "num_base_bdevs_discovered": 1, 00:16:48.319 "num_base_bdevs_operational": 2, 00:16:48.319 "base_bdevs_list": [ 00:16:48.319 { 00:16:48.319 "name": "BaseBdev1", 00:16:48.319 "uuid": "d3562751-10c9-4e82-8d2c-66a0afe5d883", 00:16:48.319 "is_configured": true, 00:16:48.319 "data_offset": 256, 00:16:48.319 "data_size": 7936 00:16:48.319 }, 00:16:48.319 { 00:16:48.319 "name": "BaseBdev2", 00:16:48.319 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:48.319 "is_configured": false, 00:16:48.319 "data_offset": 0, 00:16:48.319 "data_size": 0 00:16:48.319 } 00:16:48.319 ] 00:16:48.319 }' 00:16:48.319 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:48.319 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.577 [2024-11-20 16:04:46.657229] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:48.577 BaseBdev2 00:16:48.577 [2024-11-20 16:04:46.657584] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:16:48.577 [2024-11-20 16:04:46.657600] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:48.577 [2024-11-20 16:04:46.657835] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:16:48.577 [2024-11-20 16:04:46.657956] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:16:48.577 [2024-11-20 16:04:46.657967] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:16:48.577 [2024-11-20 16:04:46.658075] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@905 -- # local i 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.577 [ 00:16:48.577 { 00:16:48.577 "name": "BaseBdev2", 00:16:48.577 "aliases": [ 00:16:48.577 "0ce7c889-4b48-44e6-80a0-d03d0c075e35" 00:16:48.577 ], 00:16:48.577 "product_name": "Malloc disk", 00:16:48.577 "block_size": 4096, 00:16:48.577 "num_blocks": 8192, 00:16:48.577 "uuid": "0ce7c889-4b48-44e6-80a0-d03d0c075e35", 00:16:48.577 "assigned_rate_limits": { 00:16:48.577 "rw_ios_per_sec": 0, 00:16:48.577 "rw_mbytes_per_sec": 0, 00:16:48.577 "r_mbytes_per_sec": 0, 00:16:48.577 "w_mbytes_per_sec": 0 00:16:48.577 }, 00:16:48.577 "claimed": true, 00:16:48.577 "claim_type": "exclusive_write", 00:16:48.577 "zoned": false, 00:16:48.577 "supported_io_types": { 00:16:48.577 "read": true, 00:16:48.577 "write": true, 00:16:48.577 "unmap": true, 00:16:48.577 "flush": true, 00:16:48.577 "reset": true, 00:16:48.577 "nvme_admin": false, 00:16:48.577 "nvme_io": false, 00:16:48.577 "nvme_io_md": false, 00:16:48.577 "write_zeroes": true, 00:16:48.577 "zcopy": true, 00:16:48.577 "get_zone_info": false, 00:16:48.577 "zone_management": false, 00:16:48.577 "zone_append": false, 00:16:48.577 "compare": false, 00:16:48.577 "compare_and_write": false, 00:16:48.577 "abort": true, 00:16:48.577 "seek_hole": false, 00:16:48.577 "seek_data": false, 00:16:48.577 "copy": true, 00:16:48.577 "nvme_iov_md": false 00:16:48.577 }, 00:16:48.577 "memory_domains": [ 00:16:48.577 { 00:16:48.577 "dma_device_id": "system", 00:16:48.577 "dma_device_type": 1 00:16:48.577 }, 00:16:48.577 { 00:16:48.577 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:48.577 "dma_device_type": 2 00:16:48.577 } 00:16:48.577 ], 00:16:48.577 "driver_specific": {} 00:16:48.577 } 00:16:48.577 ] 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@911 -- # return 0 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.577 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:48.578 "name": "Existed_Raid", 00:16:48.578 "uuid": "b2623de6-2f65-4d81-9245-8f1c3734e778", 00:16:48.578 "strip_size_kb": 0, 00:16:48.578 "state": "online", 00:16:48.578 "raid_level": "raid1", 00:16:48.578 "superblock": true, 00:16:48.578 "num_base_bdevs": 2, 00:16:48.578 "num_base_bdevs_discovered": 2, 00:16:48.578 "num_base_bdevs_operational": 2, 00:16:48.578 "base_bdevs_list": [ 00:16:48.578 { 00:16:48.578 "name": "BaseBdev1", 00:16:48.578 "uuid": "d3562751-10c9-4e82-8d2c-66a0afe5d883", 00:16:48.578 "is_configured": true, 00:16:48.578 "data_offset": 256, 00:16:48.578 "data_size": 7936 00:16:48.578 }, 00:16:48.578 { 00:16:48.578 "name": "BaseBdev2", 00:16:48.578 "uuid": "0ce7c889-4b48-44e6-80a0-d03d0c075e35", 00:16:48.578 "is_configured": true, 00:16:48.578 "data_offset": 256, 00:16:48.578 "data_size": 7936 00:16:48.578 } 00:16:48.578 ] 00:16:48.578 }' 00:16:48.578 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:48.578 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.835 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:16:48.835 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:16:48.835 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:48.835 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:48.835 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local name 00:16:48.835 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:48.835 16:04:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:16:48.835 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.835 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.835 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:48.835 [2024-11-20 16:04:47.005566] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:48.835 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.836 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:48.836 "name": "Existed_Raid", 00:16:48.836 "aliases": [ 00:16:48.836 "b2623de6-2f65-4d81-9245-8f1c3734e778" 00:16:48.836 ], 00:16:48.836 "product_name": "Raid Volume", 00:16:48.836 "block_size": 4096, 00:16:48.836 "num_blocks": 7936, 00:16:48.836 "uuid": "b2623de6-2f65-4d81-9245-8f1c3734e778", 00:16:48.836 "assigned_rate_limits": { 00:16:48.836 "rw_ios_per_sec": 0, 00:16:48.836 "rw_mbytes_per_sec": 0, 00:16:48.836 "r_mbytes_per_sec": 0, 00:16:48.836 "w_mbytes_per_sec": 0 00:16:48.836 }, 00:16:48.836 "claimed": false, 00:16:48.836 "zoned": false, 00:16:48.836 "supported_io_types": { 00:16:48.836 "read": true, 00:16:48.836 "write": true, 00:16:48.836 "unmap": false, 00:16:48.836 "flush": false, 00:16:48.836 "reset": true, 00:16:48.836 "nvme_admin": false, 00:16:48.836 "nvme_io": false, 00:16:48.836 "nvme_io_md": false, 00:16:48.836 "write_zeroes": true, 00:16:48.836 "zcopy": false, 00:16:48.836 "get_zone_info": false, 00:16:48.836 "zone_management": false, 00:16:48.836 "zone_append": false, 00:16:48.836 "compare": false, 00:16:48.836 "compare_and_write": false, 00:16:48.836 "abort": false, 00:16:48.836 "seek_hole": false, 00:16:48.836 "seek_data": false, 00:16:48.836 "copy": false, 00:16:48.836 "nvme_iov_md": false 00:16:48.836 }, 00:16:48.836 "memory_domains": [ 00:16:48.836 { 00:16:48.836 "dma_device_id": "system", 00:16:48.836 "dma_device_type": 1 00:16:48.836 }, 00:16:48.836 { 00:16:48.836 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:48.836 "dma_device_type": 2 00:16:48.836 }, 00:16:48.836 { 00:16:48.836 "dma_device_id": "system", 00:16:48.836 "dma_device_type": 1 00:16:48.836 }, 00:16:48.836 { 00:16:48.836 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:48.836 "dma_device_type": 2 00:16:48.836 } 00:16:48.836 ], 00:16:48.836 "driver_specific": { 00:16:48.836 "raid": { 00:16:48.836 "uuid": "b2623de6-2f65-4d81-9245-8f1c3734e778", 00:16:48.836 "strip_size_kb": 0, 00:16:48.836 "state": "online", 00:16:48.836 "raid_level": "raid1", 00:16:48.836 "superblock": true, 00:16:48.836 "num_base_bdevs": 2, 00:16:48.836 "num_base_bdevs_discovered": 2, 00:16:48.836 "num_base_bdevs_operational": 2, 00:16:48.836 "base_bdevs_list": [ 00:16:48.836 { 00:16:48.836 "name": "BaseBdev1", 00:16:48.836 "uuid": "d3562751-10c9-4e82-8d2c-66a0afe5d883", 00:16:48.836 "is_configured": true, 00:16:48.836 "data_offset": 256, 00:16:48.836 "data_size": 7936 00:16:48.836 }, 00:16:48.836 { 00:16:48.836 "name": "BaseBdev2", 00:16:48.836 "uuid": "0ce7c889-4b48-44e6-80a0-d03d0c075e35", 00:16:48.836 "is_configured": true, 00:16:48.836 "data_offset": 256, 00:16:48.836 "data_size": 7936 00:16:48.836 } 00:16:48.836 ] 00:16:48.836 } 00:16:48.836 } 00:16:48.836 }' 00:16:48.836 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:48.836 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:16:48.836 BaseBdev2' 00:16:48.836 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:49.093 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:16:49.093 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:49.093 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:16:49.093 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:49.093 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.093 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.094 [2024-11-20 16:04:47.181393] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@260 -- # local expected_state 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:49.094 "name": "Existed_Raid", 00:16:49.094 "uuid": "b2623de6-2f65-4d81-9245-8f1c3734e778", 00:16:49.094 "strip_size_kb": 0, 00:16:49.094 "state": "online", 00:16:49.094 "raid_level": "raid1", 00:16:49.094 "superblock": true, 00:16:49.094 "num_base_bdevs": 2, 00:16:49.094 "num_base_bdevs_discovered": 1, 00:16:49.094 "num_base_bdevs_operational": 1, 00:16:49.094 "base_bdevs_list": [ 00:16:49.094 { 00:16:49.094 "name": null, 00:16:49.094 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:49.094 "is_configured": false, 00:16:49.094 "data_offset": 0, 00:16:49.094 "data_size": 7936 00:16:49.094 }, 00:16:49.094 { 00:16:49.094 "name": "BaseBdev2", 00:16:49.094 "uuid": "0ce7c889-4b48-44e6-80a0-d03d0c075e35", 00:16:49.094 "is_configured": true, 00:16:49.094 "data_offset": 256, 00:16:49.094 "data_size": 7936 00:16:49.094 } 00:16:49.094 ] 00:16:49.094 }' 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:49.094 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.350 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:16:49.350 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:49.350 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:49.350 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:16:49.350 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.350 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.350 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.350 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:16:49.350 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:49.351 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:16:49.351 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.351 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.351 [2024-11-20 16:04:47.571989] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:49.351 [2024-11-20 16:04:47.572071] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:49.607 [2024-11-20 16:04:47.618553] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:49.607 [2024-11-20 16:04:47.618731] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:49.607 [2024-11-20 16:04:47.618803] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@326 -- # killprocess 83525 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@954 -- # '[' -z 83525 ']' 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@958 -- # kill -0 83525 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@959 -- # uname 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83525 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83525' 00:16:49.607 killing process with pid 83525 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@973 -- # kill 83525 00:16:49.607 [2024-11-20 16:04:47.681711] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:49.607 16:04:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@978 -- # wait 83525 00:16:49.607 [2024-11-20 16:04:47.690351] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:50.213 16:04:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@328 -- # return 0 00:16:50.213 ************************************ 00:16:50.213 END TEST raid_state_function_test_sb_4k 00:16:50.213 ************************************ 00:16:50.213 00:16:50.213 real 0m3.579s 00:16:50.213 user 0m5.252s 00:16:50.213 sys 0m0.584s 00:16:50.213 16:04:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:50.213 16:04:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:50.213 16:04:48 bdev_raid -- bdev/bdev_raid.sh@998 -- # run_test raid_superblock_test_4k raid_superblock_test raid1 2 00:16:50.213 16:04:48 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:16:50.213 16:04:48 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:50.213 16:04:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:50.213 ************************************ 00:16:50.213 START TEST raid_superblock_test_4k 00:16:50.213 ************************************ 00:16:50.213 16:04:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:16:50.213 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:16:50.213 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:16:50.213 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:16:50.213 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@399 -- # local strip_size 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@412 -- # raid_pid=83762 00:16:50.214 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@413 -- # waitforlisten 83762 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@835 -- # '[' -z 83762 ']' 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:50.214 16:04:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:50.214 [2024-11-20 16:04:48.364843] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:16:50.214 [2024-11-20 16:04:48.365147] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83762 ] 00:16:50.470 [2024-11-20 16:04:48.521323] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:50.470 [2024-11-20 16:04:48.607080] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:50.470 [2024-11-20 16:04:48.717530] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:50.470 [2024-11-20 16:04:48.717563] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@868 -- # return 0 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc1 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.034 malloc1 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.034 [2024-11-20 16:04:49.243863] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:51.034 [2024-11-20 16:04:49.243918] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:51.034 [2024-11-20 16:04:49.243941] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:51.034 [2024-11-20 16:04:49.243951] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:51.034 [2024-11-20 16:04:49.245866] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:51.034 [2024-11-20 16:04:49.245896] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:51.034 pt1 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc2 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.034 malloc2 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.034 [2024-11-20 16:04:49.275516] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:51.034 [2024-11-20 16:04:49.275639] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:51.034 [2024-11-20 16:04:49.275664] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:51.034 [2024-11-20 16:04:49.275685] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:51.034 [2024-11-20 16:04:49.277392] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:51.034 [2024-11-20 16:04:49.277421] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:51.034 pt2 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.034 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.292 [2024-11-20 16:04:49.283551] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:51.292 [2024-11-20 16:04:49.285056] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:51.292 [2024-11-20 16:04:49.285185] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:16:51.292 [2024-11-20 16:04:49.285197] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:51.292 [2024-11-20 16:04:49.285391] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:16:51.292 [2024-11-20 16:04:49.285501] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:16:51.292 [2024-11-20 16:04:49.285513] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:16:51.292 [2024-11-20 16:04:49.285622] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:51.292 "name": "raid_bdev1", 00:16:51.292 "uuid": "80aed8e3-1c9a-4147-93a3-939f4e210a82", 00:16:51.292 "strip_size_kb": 0, 00:16:51.292 "state": "online", 00:16:51.292 "raid_level": "raid1", 00:16:51.292 "superblock": true, 00:16:51.292 "num_base_bdevs": 2, 00:16:51.292 "num_base_bdevs_discovered": 2, 00:16:51.292 "num_base_bdevs_operational": 2, 00:16:51.292 "base_bdevs_list": [ 00:16:51.292 { 00:16:51.292 "name": "pt1", 00:16:51.292 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:51.292 "is_configured": true, 00:16:51.292 "data_offset": 256, 00:16:51.292 "data_size": 7936 00:16:51.292 }, 00:16:51.292 { 00:16:51.292 "name": "pt2", 00:16:51.292 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:51.292 "is_configured": true, 00:16:51.292 "data_offset": 256, 00:16:51.292 "data_size": 7936 00:16:51.292 } 00:16:51.292 ] 00:16:51.292 }' 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:51.292 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.551 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:16:51.551 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:51.551 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:51.551 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:51.551 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:16:51.551 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:51.551 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:51.551 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:51.551 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.551 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.551 [2024-11-20 16:04:49.607853] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:51.551 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.551 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:51.551 "name": "raid_bdev1", 00:16:51.551 "aliases": [ 00:16:51.551 "80aed8e3-1c9a-4147-93a3-939f4e210a82" 00:16:51.551 ], 00:16:51.551 "product_name": "Raid Volume", 00:16:51.551 "block_size": 4096, 00:16:51.551 "num_blocks": 7936, 00:16:51.551 "uuid": "80aed8e3-1c9a-4147-93a3-939f4e210a82", 00:16:51.551 "assigned_rate_limits": { 00:16:51.551 "rw_ios_per_sec": 0, 00:16:51.551 "rw_mbytes_per_sec": 0, 00:16:51.551 "r_mbytes_per_sec": 0, 00:16:51.551 "w_mbytes_per_sec": 0 00:16:51.551 }, 00:16:51.551 "claimed": false, 00:16:51.551 "zoned": false, 00:16:51.551 "supported_io_types": { 00:16:51.551 "read": true, 00:16:51.551 "write": true, 00:16:51.551 "unmap": false, 00:16:51.551 "flush": false, 00:16:51.551 "reset": true, 00:16:51.551 "nvme_admin": false, 00:16:51.551 "nvme_io": false, 00:16:51.551 "nvme_io_md": false, 00:16:51.551 "write_zeroes": true, 00:16:51.551 "zcopy": false, 00:16:51.551 "get_zone_info": false, 00:16:51.551 "zone_management": false, 00:16:51.551 "zone_append": false, 00:16:51.551 "compare": false, 00:16:51.551 "compare_and_write": false, 00:16:51.551 "abort": false, 00:16:51.551 "seek_hole": false, 00:16:51.551 "seek_data": false, 00:16:51.551 "copy": false, 00:16:51.551 "nvme_iov_md": false 00:16:51.551 }, 00:16:51.551 "memory_domains": [ 00:16:51.551 { 00:16:51.551 "dma_device_id": "system", 00:16:51.551 "dma_device_type": 1 00:16:51.551 }, 00:16:51.551 { 00:16:51.551 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:51.551 "dma_device_type": 2 00:16:51.551 }, 00:16:51.551 { 00:16:51.551 "dma_device_id": "system", 00:16:51.551 "dma_device_type": 1 00:16:51.551 }, 00:16:51.551 { 00:16:51.551 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:51.551 "dma_device_type": 2 00:16:51.551 } 00:16:51.551 ], 00:16:51.551 "driver_specific": { 00:16:51.551 "raid": { 00:16:51.551 "uuid": "80aed8e3-1c9a-4147-93a3-939f4e210a82", 00:16:51.551 "strip_size_kb": 0, 00:16:51.551 "state": "online", 00:16:51.551 "raid_level": "raid1", 00:16:51.551 "superblock": true, 00:16:51.551 "num_base_bdevs": 2, 00:16:51.551 "num_base_bdevs_discovered": 2, 00:16:51.551 "num_base_bdevs_operational": 2, 00:16:51.551 "base_bdevs_list": [ 00:16:51.551 { 00:16:51.551 "name": "pt1", 00:16:51.551 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:51.551 "is_configured": true, 00:16:51.551 "data_offset": 256, 00:16:51.551 "data_size": 7936 00:16:51.552 }, 00:16:51.552 { 00:16:51.552 "name": "pt2", 00:16:51.552 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:51.552 "is_configured": true, 00:16:51.552 "data_offset": 256, 00:16:51.552 "data_size": 7936 00:16:51.552 } 00:16:51.552 ] 00:16:51.552 } 00:16:51.552 } 00:16:51.552 }' 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:51.552 pt2' 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.552 [2024-11-20 16:04:49.775861] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=80aed8e3-1c9a-4147-93a3-939f4e210a82 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@436 -- # '[' -z 80aed8e3-1c9a-4147-93a3-939f4e210a82 ']' 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.552 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.552 [2024-11-20 16:04:49.799602] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:51.552 [2024-11-20 16:04:49.799702] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:51.552 [2024-11-20 16:04:49.799827] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:51.552 [2024-11-20 16:04:49.799914] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:51.552 [2024-11-20 16:04:49.799969] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:16:51.809 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.809 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:16:51.809 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:51.809 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.809 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.809 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.809 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:16:51.809 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:16:51.809 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@652 -- # local es=0 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.810 [2024-11-20 16:04:49.891659] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:16:51.810 [2024-11-20 16:04:49.893222] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:16:51.810 [2024-11-20 16:04:49.893279] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:16:51.810 [2024-11-20 16:04:49.893323] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:16:51.810 [2024-11-20 16:04:49.893335] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:51.810 [2024-11-20 16:04:49.893344] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:16:51.810 request: 00:16:51.810 { 00:16:51.810 "name": "raid_bdev1", 00:16:51.810 "raid_level": "raid1", 00:16:51.810 "base_bdevs": [ 00:16:51.810 "malloc1", 00:16:51.810 "malloc2" 00:16:51.810 ], 00:16:51.810 "superblock": false, 00:16:51.810 "method": "bdev_raid_create", 00:16:51.810 "req_id": 1 00:16:51.810 } 00:16:51.810 Got JSON-RPC error response 00:16:51.810 response: 00:16:51.810 { 00:16:51.810 "code": -17, 00:16:51.810 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:16:51.810 } 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@655 -- # es=1 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.810 [2024-11-20 16:04:49.943658] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:51.810 [2024-11-20 16:04:49.943806] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:51.810 [2024-11-20 16:04:49.943839] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:16:51.810 [2024-11-20 16:04:49.943885] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:51.810 [2024-11-20 16:04:49.945714] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:51.810 [2024-11-20 16:04:49.945808] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:51.810 [2024-11-20 16:04:49.945954] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:51.810 [2024-11-20 16:04:49.946018] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:51.810 pt1 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.810 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:51.810 "name": "raid_bdev1", 00:16:51.810 "uuid": "80aed8e3-1c9a-4147-93a3-939f4e210a82", 00:16:51.810 "strip_size_kb": 0, 00:16:51.810 "state": "configuring", 00:16:51.810 "raid_level": "raid1", 00:16:51.810 "superblock": true, 00:16:51.810 "num_base_bdevs": 2, 00:16:51.810 "num_base_bdevs_discovered": 1, 00:16:51.810 "num_base_bdevs_operational": 2, 00:16:51.810 "base_bdevs_list": [ 00:16:51.810 { 00:16:51.810 "name": "pt1", 00:16:51.810 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:51.810 "is_configured": true, 00:16:51.810 "data_offset": 256, 00:16:51.810 "data_size": 7936 00:16:51.810 }, 00:16:51.810 { 00:16:51.810 "name": null, 00:16:51.810 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:51.810 "is_configured": false, 00:16:51.810 "data_offset": 256, 00:16:51.811 "data_size": 7936 00:16:51.811 } 00:16:51.811 ] 00:16:51.811 }' 00:16:51.811 16:04:49 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:51.811 16:04:49 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.068 [2024-11-20 16:04:50.259719] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:52.068 [2024-11-20 16:04:50.259769] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:52.068 [2024-11-20 16:04:50.259783] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:16:52.068 [2024-11-20 16:04:50.259792] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:52.068 [2024-11-20 16:04:50.260129] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:52.068 [2024-11-20 16:04:50.260141] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:52.068 [2024-11-20 16:04:50.260200] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:52.068 [2024-11-20 16:04:50.260219] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:52.068 [2024-11-20 16:04:50.260306] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:16:52.068 [2024-11-20 16:04:50.260314] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:52.068 [2024-11-20 16:04:50.260499] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:16:52.068 [2024-11-20 16:04:50.260605] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:16:52.068 [2024-11-20 16:04:50.260611] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:16:52.068 [2024-11-20 16:04:50.260728] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:52.068 pt2 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:52.068 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:52.068 "name": "raid_bdev1", 00:16:52.069 "uuid": "80aed8e3-1c9a-4147-93a3-939f4e210a82", 00:16:52.069 "strip_size_kb": 0, 00:16:52.069 "state": "online", 00:16:52.069 "raid_level": "raid1", 00:16:52.069 "superblock": true, 00:16:52.069 "num_base_bdevs": 2, 00:16:52.069 "num_base_bdevs_discovered": 2, 00:16:52.069 "num_base_bdevs_operational": 2, 00:16:52.069 "base_bdevs_list": [ 00:16:52.069 { 00:16:52.069 "name": "pt1", 00:16:52.069 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:52.069 "is_configured": true, 00:16:52.069 "data_offset": 256, 00:16:52.069 "data_size": 7936 00:16:52.069 }, 00:16:52.069 { 00:16:52.069 "name": "pt2", 00:16:52.069 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:52.069 "is_configured": true, 00:16:52.069 "data_offset": 256, 00:16:52.069 "data_size": 7936 00:16:52.069 } 00:16:52.069 ] 00:16:52.069 }' 00:16:52.069 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:52.069 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:52.634 [2024-11-20 16:04:50.620007] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:52.634 "name": "raid_bdev1", 00:16:52.634 "aliases": [ 00:16:52.634 "80aed8e3-1c9a-4147-93a3-939f4e210a82" 00:16:52.634 ], 00:16:52.634 "product_name": "Raid Volume", 00:16:52.634 "block_size": 4096, 00:16:52.634 "num_blocks": 7936, 00:16:52.634 "uuid": "80aed8e3-1c9a-4147-93a3-939f4e210a82", 00:16:52.634 "assigned_rate_limits": { 00:16:52.634 "rw_ios_per_sec": 0, 00:16:52.634 "rw_mbytes_per_sec": 0, 00:16:52.634 "r_mbytes_per_sec": 0, 00:16:52.634 "w_mbytes_per_sec": 0 00:16:52.634 }, 00:16:52.634 "claimed": false, 00:16:52.634 "zoned": false, 00:16:52.634 "supported_io_types": { 00:16:52.634 "read": true, 00:16:52.634 "write": true, 00:16:52.634 "unmap": false, 00:16:52.634 "flush": false, 00:16:52.634 "reset": true, 00:16:52.634 "nvme_admin": false, 00:16:52.634 "nvme_io": false, 00:16:52.634 "nvme_io_md": false, 00:16:52.634 "write_zeroes": true, 00:16:52.634 "zcopy": false, 00:16:52.634 "get_zone_info": false, 00:16:52.634 "zone_management": false, 00:16:52.634 "zone_append": false, 00:16:52.634 "compare": false, 00:16:52.634 "compare_and_write": false, 00:16:52.634 "abort": false, 00:16:52.634 "seek_hole": false, 00:16:52.634 "seek_data": false, 00:16:52.634 "copy": false, 00:16:52.634 "nvme_iov_md": false 00:16:52.634 }, 00:16:52.634 "memory_domains": [ 00:16:52.634 { 00:16:52.634 "dma_device_id": "system", 00:16:52.634 "dma_device_type": 1 00:16:52.634 }, 00:16:52.634 { 00:16:52.634 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:52.634 "dma_device_type": 2 00:16:52.634 }, 00:16:52.634 { 00:16:52.634 "dma_device_id": "system", 00:16:52.634 "dma_device_type": 1 00:16:52.634 }, 00:16:52.634 { 00:16:52.634 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:52.634 "dma_device_type": 2 00:16:52.634 } 00:16:52.634 ], 00:16:52.634 "driver_specific": { 00:16:52.634 "raid": { 00:16:52.634 "uuid": "80aed8e3-1c9a-4147-93a3-939f4e210a82", 00:16:52.634 "strip_size_kb": 0, 00:16:52.634 "state": "online", 00:16:52.634 "raid_level": "raid1", 00:16:52.634 "superblock": true, 00:16:52.634 "num_base_bdevs": 2, 00:16:52.634 "num_base_bdevs_discovered": 2, 00:16:52.634 "num_base_bdevs_operational": 2, 00:16:52.634 "base_bdevs_list": [ 00:16:52.634 { 00:16:52.634 "name": "pt1", 00:16:52.634 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:52.634 "is_configured": true, 00:16:52.634 "data_offset": 256, 00:16:52.634 "data_size": 7936 00:16:52.634 }, 00:16:52.634 { 00:16:52.634 "name": "pt2", 00:16:52.634 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:52.634 "is_configured": true, 00:16:52.634 "data_offset": 256, 00:16:52.634 "data_size": 7936 00:16:52.634 } 00:16:52.634 ] 00:16:52.634 } 00:16:52.634 } 00:16:52.634 }' 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:52.634 pt2' 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:16:52.634 [2024-11-20 16:04:50.788036] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # '[' 80aed8e3-1c9a-4147-93a3-939f4e210a82 '!=' 80aed8e3-1c9a-4147-93a3-939f4e210a82 ']' 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:52.634 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.635 [2024-11-20 16:04:50.815846] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:52.635 "name": "raid_bdev1", 00:16:52.635 "uuid": "80aed8e3-1c9a-4147-93a3-939f4e210a82", 00:16:52.635 "strip_size_kb": 0, 00:16:52.635 "state": "online", 00:16:52.635 "raid_level": "raid1", 00:16:52.635 "superblock": true, 00:16:52.635 "num_base_bdevs": 2, 00:16:52.635 "num_base_bdevs_discovered": 1, 00:16:52.635 "num_base_bdevs_operational": 1, 00:16:52.635 "base_bdevs_list": [ 00:16:52.635 { 00:16:52.635 "name": null, 00:16:52.635 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:52.635 "is_configured": false, 00:16:52.635 "data_offset": 0, 00:16:52.635 "data_size": 7936 00:16:52.635 }, 00:16:52.635 { 00:16:52.635 "name": "pt2", 00:16:52.635 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:52.635 "is_configured": true, 00:16:52.635 "data_offset": 256, 00:16:52.635 "data_size": 7936 00:16:52.635 } 00:16:52.635 ] 00:16:52.635 }' 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:52.635 16:04:50 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.893 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:52.893 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.893 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.893 [2024-11-20 16:04:51.139883] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:52.893 [2024-11-20 16:04:51.139902] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:52.893 [2024-11-20 16:04:51.139955] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:52.893 [2024-11-20 16:04:51.139991] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:52.893 [2024-11-20 16:04:51.140000] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@519 -- # i=1 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.151 [2024-11-20 16:04:51.191881] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:53.151 [2024-11-20 16:04:51.191921] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:53.151 [2024-11-20 16:04:51.191934] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:16:53.151 [2024-11-20 16:04:51.191943] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:53.151 [2024-11-20 16:04:51.193744] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:53.151 [2024-11-20 16:04:51.193860] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:53.151 [2024-11-20 16:04:51.193927] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:53.151 [2024-11-20 16:04:51.193965] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:53.151 [2024-11-20 16:04:51.194041] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:16:53.151 [2024-11-20 16:04:51.194051] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:53.151 [2024-11-20 16:04:51.194239] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:16:53.151 [2024-11-20 16:04:51.194346] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:16:53.151 [2024-11-20 16:04:51.194353] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:16:53.151 [2024-11-20 16:04:51.194454] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:53.151 pt2 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:53.151 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.152 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.152 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.152 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:53.152 "name": "raid_bdev1", 00:16:53.152 "uuid": "80aed8e3-1c9a-4147-93a3-939f4e210a82", 00:16:53.152 "strip_size_kb": 0, 00:16:53.152 "state": "online", 00:16:53.152 "raid_level": "raid1", 00:16:53.152 "superblock": true, 00:16:53.152 "num_base_bdevs": 2, 00:16:53.152 "num_base_bdevs_discovered": 1, 00:16:53.152 "num_base_bdevs_operational": 1, 00:16:53.152 "base_bdevs_list": [ 00:16:53.152 { 00:16:53.152 "name": null, 00:16:53.152 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:53.152 "is_configured": false, 00:16:53.152 "data_offset": 256, 00:16:53.152 "data_size": 7936 00:16:53.152 }, 00:16:53.152 { 00:16:53.152 "name": "pt2", 00:16:53.152 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:53.152 "is_configured": true, 00:16:53.152 "data_offset": 256, 00:16:53.152 "data_size": 7936 00:16:53.152 } 00:16:53.152 ] 00:16:53.152 }' 00:16:53.152 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:53.152 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.410 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:53.410 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.410 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.410 [2024-11-20 16:04:51.511935] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:53.410 [2024-11-20 16:04:51.511961] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:53.410 [2024-11-20 16:04:51.512015] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:53.410 [2024-11-20 16:04:51.512054] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:53.410 [2024-11-20 16:04:51.512062] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:16:53.410 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.410 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:53.410 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.410 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.410 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:16:53.410 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.410 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:16:53.410 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.411 [2024-11-20 16:04:51.551959] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:53.411 [2024-11-20 16:04:51.552096] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:53.411 [2024-11-20 16:04:51.552117] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:16:53.411 [2024-11-20 16:04:51.552125] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:53.411 [2024-11-20 16:04:51.553963] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:53.411 [2024-11-20 16:04:51.553990] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:53.411 [2024-11-20 16:04:51.554056] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:53.411 [2024-11-20 16:04:51.554091] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:53.411 [2024-11-20 16:04:51.554195] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:16:53.411 [2024-11-20 16:04:51.554203] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:53.411 [2024-11-20 16:04:51.554216] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:16:53.411 [2024-11-20 16:04:51.554254] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:53.411 [2024-11-20 16:04:51.554309] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:16:53.411 [2024-11-20 16:04:51.554316] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:53.411 [2024-11-20 16:04:51.554517] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:16:53.411 [2024-11-20 16:04:51.554618] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:16:53.411 [2024-11-20 16:04:51.554631] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:16:53.411 [2024-11-20 16:04:51.554758] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:53.411 pt1 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:53.411 "name": "raid_bdev1", 00:16:53.411 "uuid": "80aed8e3-1c9a-4147-93a3-939f4e210a82", 00:16:53.411 "strip_size_kb": 0, 00:16:53.411 "state": "online", 00:16:53.411 "raid_level": "raid1", 00:16:53.411 "superblock": true, 00:16:53.411 "num_base_bdevs": 2, 00:16:53.411 "num_base_bdevs_discovered": 1, 00:16:53.411 "num_base_bdevs_operational": 1, 00:16:53.411 "base_bdevs_list": [ 00:16:53.411 { 00:16:53.411 "name": null, 00:16:53.411 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:53.411 "is_configured": false, 00:16:53.411 "data_offset": 256, 00:16:53.411 "data_size": 7936 00:16:53.411 }, 00:16:53.411 { 00:16:53.411 "name": "pt2", 00:16:53.411 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:53.411 "is_configured": true, 00:16:53.411 "data_offset": 256, 00:16:53.411 "data_size": 7936 00:16:53.411 } 00:16:53.411 ] 00:16:53.411 }' 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:53.411 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.669 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:16:53.669 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:16:53.670 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.670 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.670 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.928 [2024-11-20 16:04:51.932224] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # '[' 80aed8e3-1c9a-4147-93a3-939f4e210a82 '!=' 80aed8e3-1c9a-4147-93a3-939f4e210a82 ']' 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@563 -- # killprocess 83762 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@954 -- # '[' -z 83762 ']' 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@958 -- # kill -0 83762 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@959 -- # uname 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83762 00:16:53.928 killing process with pid 83762 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83762' 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@973 -- # kill 83762 00:16:53.928 [2024-11-20 16:04:51.984168] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:53.928 [2024-11-20 16:04:51.984239] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:53.928 16:04:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@978 -- # wait 83762 00:16:53.928 [2024-11-20 16:04:51.984276] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:53.928 [2024-11-20 16:04:51.984290] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:16:53.928 [2024-11-20 16:04:52.085058] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:54.543 ************************************ 00:16:54.543 END TEST raid_superblock_test_4k 00:16:54.543 ************************************ 00:16:54.543 16:04:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@565 -- # return 0 00:16:54.543 00:16:54.543 real 0m4.365s 00:16:54.543 user 0m6.748s 00:16:54.543 sys 0m0.704s 00:16:54.543 16:04:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:54.543 16:04:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:54.543 16:04:52 bdev_raid -- bdev/bdev_raid.sh@999 -- # '[' true = true ']' 00:16:54.543 16:04:52 bdev_raid -- bdev/bdev_raid.sh@1000 -- # run_test raid_rebuild_test_sb_4k raid_rebuild_test raid1 2 true false true 00:16:54.543 16:04:52 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:16:54.543 16:04:52 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:54.543 16:04:52 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:54.543 ************************************ 00:16:54.543 START TEST raid_rebuild_test_sb_4k 00:16:54.543 ************************************ 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # local verify=true 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:54.543 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # local strip_size 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@577 -- # local create_arg 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@579 -- # local data_offset 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@597 -- # raid_pid=84068 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@598 -- # waitforlisten 84068 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@835 -- # '[' -z 84068 ']' 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:54.543 16:04:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:54.543 I/O size of 3145728 is greater than zero copy threshold (65536). 00:16:54.543 Zero copy mechanism will not be used. 00:16:54.543 [2024-11-20 16:04:52.778409] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:16:54.543 [2024-11-20 16:04:52.778519] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid84068 ] 00:16:54.801 [2024-11-20 16:04:52.935928] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:54.801 [2024-11-20 16:04:53.020267] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:55.060 [2024-11-20 16:04:53.157074] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:55.060 [2024-11-20 16:04:53.157318] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # return 0 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1_malloc 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:55.626 BaseBdev1_malloc 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:55.626 [2024-11-20 16:04:53.614896] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:55.626 [2024-11-20 16:04:53.615074] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:55.626 [2024-11-20 16:04:53.615098] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:55.626 [2024-11-20 16:04:53.615108] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:55.626 [2024-11-20 16:04:53.616877] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:55.626 [2024-11-20 16:04:53.616909] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:55.626 BaseBdev1 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2_malloc 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:55.626 BaseBdev2_malloc 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:55.626 [2024-11-20 16:04:53.646388] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:16:55.626 [2024-11-20 16:04:53.646541] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:55.626 [2024-11-20 16:04:53.646565] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:55.626 [2024-11-20 16:04:53.646573] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:55.626 [2024-11-20 16:04:53.648303] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:55.626 [2024-11-20 16:04:53.648330] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:55.626 BaseBdev2 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -b spare_malloc 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:55.626 spare_malloc 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:55.626 spare_delay 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:55.626 [2024-11-20 16:04:53.700586] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:55.626 [2024-11-20 16:04:53.700637] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:55.626 [2024-11-20 16:04:53.700652] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:16:55.626 [2024-11-20 16:04:53.700661] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:55.626 [2024-11-20 16:04:53.702446] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:55.626 [2024-11-20 16:04:53.702600] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:55.626 spare 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.626 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:55.626 [2024-11-20 16:04:53.708635] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:55.626 [2024-11-20 16:04:53.710154] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:55.626 [2024-11-20 16:04:53.710289] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:16:55.626 [2024-11-20 16:04:53.710301] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:55.626 [2024-11-20 16:04:53.710509] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:16:55.626 [2024-11-20 16:04:53.710630] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:16:55.627 [2024-11-20 16:04:53.710637] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:16:55.627 [2024-11-20 16:04:53.710764] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:55.627 "name": "raid_bdev1", 00:16:55.627 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:16:55.627 "strip_size_kb": 0, 00:16:55.627 "state": "online", 00:16:55.627 "raid_level": "raid1", 00:16:55.627 "superblock": true, 00:16:55.627 "num_base_bdevs": 2, 00:16:55.627 "num_base_bdevs_discovered": 2, 00:16:55.627 "num_base_bdevs_operational": 2, 00:16:55.627 "base_bdevs_list": [ 00:16:55.627 { 00:16:55.627 "name": "BaseBdev1", 00:16:55.627 "uuid": "9f776538-9768-5831-94dc-07caf186e72c", 00:16:55.627 "is_configured": true, 00:16:55.627 "data_offset": 256, 00:16:55.627 "data_size": 7936 00:16:55.627 }, 00:16:55.627 { 00:16:55.627 "name": "BaseBdev2", 00:16:55.627 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:16:55.627 "is_configured": true, 00:16:55.627 "data_offset": 256, 00:16:55.627 "data_size": 7936 00:16:55.627 } 00:16:55.627 ] 00:16:55.627 }' 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:55.627 16:04:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:55.885 [2024-11-20 16:04:54.028941] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:55.885 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:16:56.143 [2024-11-20 16:04:54.264793] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:16:56.143 /dev/nbd0 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:56.143 1+0 records in 00:16:56.143 1+0 records out 00:16:56.143 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000209391 s, 19.6 MB/s 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:16:56.143 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:16:56.706 7936+0 records in 00:16:56.706 7936+0 records out 00:16:56.706 32505856 bytes (33 MB, 31 MiB) copied, 0.580996 s, 55.9 MB/s 00:16:56.706 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:16:56.707 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:56.707 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:56.707 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:56.707 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:16:56.707 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:56.707 16:04:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:56.965 [2024-11-20 16:04:55.069222] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:56.965 [2024-11-20 16:04:55.098171] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:56.965 "name": "raid_bdev1", 00:16:56.965 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:16:56.965 "strip_size_kb": 0, 00:16:56.965 "state": "online", 00:16:56.965 "raid_level": "raid1", 00:16:56.965 "superblock": true, 00:16:56.965 "num_base_bdevs": 2, 00:16:56.965 "num_base_bdevs_discovered": 1, 00:16:56.965 "num_base_bdevs_operational": 1, 00:16:56.965 "base_bdevs_list": [ 00:16:56.965 { 00:16:56.965 "name": null, 00:16:56.965 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:56.965 "is_configured": false, 00:16:56.965 "data_offset": 0, 00:16:56.965 "data_size": 7936 00:16:56.965 }, 00:16:56.965 { 00:16:56.965 "name": "BaseBdev2", 00:16:56.965 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:16:56.965 "is_configured": true, 00:16:56.965 "data_offset": 256, 00:16:56.965 "data_size": 7936 00:16:56.965 } 00:16:56.965 ] 00:16:56.965 }' 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:56.965 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:57.223 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:57.223 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:57.223 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:57.223 [2024-11-20 16:04:55.398252] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:57.223 [2024-11-20 16:04:55.408075] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d260 00:16:57.223 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:57.223 16:04:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:57.223 [2024-11-20 16:04:55.409647] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:58.194 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:58.194 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:58.194 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:58.194 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:58.194 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:58.194 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:58.194 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:58.194 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:58.194 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:58.194 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:58.453 "name": "raid_bdev1", 00:16:58.453 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:16:58.453 "strip_size_kb": 0, 00:16:58.453 "state": "online", 00:16:58.453 "raid_level": "raid1", 00:16:58.453 "superblock": true, 00:16:58.453 "num_base_bdevs": 2, 00:16:58.453 "num_base_bdevs_discovered": 2, 00:16:58.453 "num_base_bdevs_operational": 2, 00:16:58.453 "process": { 00:16:58.453 "type": "rebuild", 00:16:58.453 "target": "spare", 00:16:58.453 "progress": { 00:16:58.453 "blocks": 2560, 00:16:58.453 "percent": 32 00:16:58.453 } 00:16:58.453 }, 00:16:58.453 "base_bdevs_list": [ 00:16:58.453 { 00:16:58.453 "name": "spare", 00:16:58.453 "uuid": "d412ac6c-c84c-55ed-a564-acfe470c4a64", 00:16:58.453 "is_configured": true, 00:16:58.453 "data_offset": 256, 00:16:58.453 "data_size": 7936 00:16:58.453 }, 00:16:58.453 { 00:16:58.453 "name": "BaseBdev2", 00:16:58.453 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:16:58.453 "is_configured": true, 00:16:58.453 "data_offset": 256, 00:16:58.453 "data_size": 7936 00:16:58.453 } 00:16:58.453 ] 00:16:58.453 }' 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:58.453 [2024-11-20 16:04:56.507584] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:58.453 [2024-11-20 16:04:56.514777] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:58.453 [2024-11-20 16:04:56.514827] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:58.453 [2024-11-20 16:04:56.514838] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:58.453 [2024-11-20 16:04:56.514848] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:58.453 "name": "raid_bdev1", 00:16:58.453 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:16:58.453 "strip_size_kb": 0, 00:16:58.453 "state": "online", 00:16:58.453 "raid_level": "raid1", 00:16:58.453 "superblock": true, 00:16:58.453 "num_base_bdevs": 2, 00:16:58.453 "num_base_bdevs_discovered": 1, 00:16:58.453 "num_base_bdevs_operational": 1, 00:16:58.453 "base_bdevs_list": [ 00:16:58.453 { 00:16:58.453 "name": null, 00:16:58.453 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:58.453 "is_configured": false, 00:16:58.453 "data_offset": 0, 00:16:58.453 "data_size": 7936 00:16:58.453 }, 00:16:58.453 { 00:16:58.453 "name": "BaseBdev2", 00:16:58.453 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:16:58.453 "is_configured": true, 00:16:58.453 "data_offset": 256, 00:16:58.453 "data_size": 7936 00:16:58.453 } 00:16:58.453 ] 00:16:58.453 }' 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:58.453 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:58.712 "name": "raid_bdev1", 00:16:58.712 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:16:58.712 "strip_size_kb": 0, 00:16:58.712 "state": "online", 00:16:58.712 "raid_level": "raid1", 00:16:58.712 "superblock": true, 00:16:58.712 "num_base_bdevs": 2, 00:16:58.712 "num_base_bdevs_discovered": 1, 00:16:58.712 "num_base_bdevs_operational": 1, 00:16:58.712 "base_bdevs_list": [ 00:16:58.712 { 00:16:58.712 "name": null, 00:16:58.712 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:58.712 "is_configured": false, 00:16:58.712 "data_offset": 0, 00:16:58.712 "data_size": 7936 00:16:58.712 }, 00:16:58.712 { 00:16:58.712 "name": "BaseBdev2", 00:16:58.712 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:16:58.712 "is_configured": true, 00:16:58.712 "data_offset": 256, 00:16:58.712 "data_size": 7936 00:16:58.712 } 00:16:58.712 ] 00:16:58.712 }' 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:58.712 [2024-11-20 16:04:56.933319] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:58.712 [2024-11-20 16:04:56.942645] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d330 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:58.712 16:04:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:58.712 [2024-11-20 16:04:56.944252] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:00.085 16:04:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:00.085 16:04:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:00.085 16:04:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:00.085 16:04:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:00.085 16:04:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:00.085 16:04:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:00.085 16:04:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:00.085 16:04:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:00.085 16:04:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:00.085 16:04:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:00.085 16:04:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:00.085 "name": "raid_bdev1", 00:17:00.085 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:00.085 "strip_size_kb": 0, 00:17:00.085 "state": "online", 00:17:00.085 "raid_level": "raid1", 00:17:00.085 "superblock": true, 00:17:00.085 "num_base_bdevs": 2, 00:17:00.085 "num_base_bdevs_discovered": 2, 00:17:00.085 "num_base_bdevs_operational": 2, 00:17:00.085 "process": { 00:17:00.085 "type": "rebuild", 00:17:00.085 "target": "spare", 00:17:00.085 "progress": { 00:17:00.085 "blocks": 2560, 00:17:00.085 "percent": 32 00:17:00.085 } 00:17:00.085 }, 00:17:00.085 "base_bdevs_list": [ 00:17:00.085 { 00:17:00.085 "name": "spare", 00:17:00.085 "uuid": "d412ac6c-c84c-55ed-a564-acfe470c4a64", 00:17:00.085 "is_configured": true, 00:17:00.085 "data_offset": 256, 00:17:00.085 "data_size": 7936 00:17:00.085 }, 00:17:00.085 { 00:17:00.085 "name": "BaseBdev2", 00:17:00.085 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:00.085 "is_configured": true, 00:17:00.085 "data_offset": 256, 00:17:00.085 "data_size": 7936 00:17:00.085 } 00:17:00.085 ] 00:17:00.085 }' 00:17:00.085 16:04:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:00.085 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:00.085 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:00.085 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:00.085 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:17:00.085 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:17:00.085 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:17:00.085 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:17:00.085 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@706 -- # local timeout=552 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:00.086 "name": "raid_bdev1", 00:17:00.086 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:00.086 "strip_size_kb": 0, 00:17:00.086 "state": "online", 00:17:00.086 "raid_level": "raid1", 00:17:00.086 "superblock": true, 00:17:00.086 "num_base_bdevs": 2, 00:17:00.086 "num_base_bdevs_discovered": 2, 00:17:00.086 "num_base_bdevs_operational": 2, 00:17:00.086 "process": { 00:17:00.086 "type": "rebuild", 00:17:00.086 "target": "spare", 00:17:00.086 "progress": { 00:17:00.086 "blocks": 2816, 00:17:00.086 "percent": 35 00:17:00.086 } 00:17:00.086 }, 00:17:00.086 "base_bdevs_list": [ 00:17:00.086 { 00:17:00.086 "name": "spare", 00:17:00.086 "uuid": "d412ac6c-c84c-55ed-a564-acfe470c4a64", 00:17:00.086 "is_configured": true, 00:17:00.086 "data_offset": 256, 00:17:00.086 "data_size": 7936 00:17:00.086 }, 00:17:00.086 { 00:17:00.086 "name": "BaseBdev2", 00:17:00.086 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:00.086 "is_configured": true, 00:17:00.086 "data_offset": 256, 00:17:00.086 "data_size": 7936 00:17:00.086 } 00:17:00.086 ] 00:17:00.086 }' 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:00.086 16:04:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:17:01.120 16:04:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:01.120 16:04:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:01.120 16:04:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:01.120 16:04:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:01.120 16:04:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:01.120 16:04:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:01.120 16:04:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:01.120 16:04:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:01.120 16:04:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:01.120 16:04:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:01.120 16:04:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:01.120 16:04:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:01.120 "name": "raid_bdev1", 00:17:01.120 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:01.120 "strip_size_kb": 0, 00:17:01.120 "state": "online", 00:17:01.120 "raid_level": "raid1", 00:17:01.120 "superblock": true, 00:17:01.120 "num_base_bdevs": 2, 00:17:01.120 "num_base_bdevs_discovered": 2, 00:17:01.120 "num_base_bdevs_operational": 2, 00:17:01.120 "process": { 00:17:01.120 "type": "rebuild", 00:17:01.120 "target": "spare", 00:17:01.120 "progress": { 00:17:01.120 "blocks": 5376, 00:17:01.120 "percent": 67 00:17:01.120 } 00:17:01.120 }, 00:17:01.120 "base_bdevs_list": [ 00:17:01.120 { 00:17:01.120 "name": "spare", 00:17:01.120 "uuid": "d412ac6c-c84c-55ed-a564-acfe470c4a64", 00:17:01.120 "is_configured": true, 00:17:01.120 "data_offset": 256, 00:17:01.120 "data_size": 7936 00:17:01.120 }, 00:17:01.120 { 00:17:01.120 "name": "BaseBdev2", 00:17:01.120 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:01.120 "is_configured": true, 00:17:01.120 "data_offset": 256, 00:17:01.120 "data_size": 7936 00:17:01.120 } 00:17:01.120 ] 00:17:01.120 }' 00:17:01.120 16:04:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:01.120 16:04:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:01.120 16:04:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:01.120 16:04:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:01.120 16:04:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:17:02.055 [2024-11-20 16:05:00.058192] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:17:02.055 [2024-11-20 16:05:00.058259] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:17:02.055 [2024-11-20 16:05:00.058361] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:02.055 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:02.055 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:02.055 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:02.055 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:02.055 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:02.055 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:02.055 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:02.055 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:02.055 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:02.055 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:02.055 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:02.055 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:02.055 "name": "raid_bdev1", 00:17:02.055 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:02.055 "strip_size_kb": 0, 00:17:02.055 "state": "online", 00:17:02.055 "raid_level": "raid1", 00:17:02.055 "superblock": true, 00:17:02.055 "num_base_bdevs": 2, 00:17:02.055 "num_base_bdevs_discovered": 2, 00:17:02.055 "num_base_bdevs_operational": 2, 00:17:02.055 "base_bdevs_list": [ 00:17:02.055 { 00:17:02.055 "name": "spare", 00:17:02.055 "uuid": "d412ac6c-c84c-55ed-a564-acfe470c4a64", 00:17:02.055 "is_configured": true, 00:17:02.055 "data_offset": 256, 00:17:02.055 "data_size": 7936 00:17:02.055 }, 00:17:02.055 { 00:17:02.055 "name": "BaseBdev2", 00:17:02.055 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:02.055 "is_configured": true, 00:17:02.055 "data_offset": 256, 00:17:02.055 "data_size": 7936 00:17:02.055 } 00:17:02.055 ] 00:17:02.055 }' 00:17:02.055 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@709 -- # break 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:02.317 "name": "raid_bdev1", 00:17:02.317 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:02.317 "strip_size_kb": 0, 00:17:02.317 "state": "online", 00:17:02.317 "raid_level": "raid1", 00:17:02.317 "superblock": true, 00:17:02.317 "num_base_bdevs": 2, 00:17:02.317 "num_base_bdevs_discovered": 2, 00:17:02.317 "num_base_bdevs_operational": 2, 00:17:02.317 "base_bdevs_list": [ 00:17:02.317 { 00:17:02.317 "name": "spare", 00:17:02.317 "uuid": "d412ac6c-c84c-55ed-a564-acfe470c4a64", 00:17:02.317 "is_configured": true, 00:17:02.317 "data_offset": 256, 00:17:02.317 "data_size": 7936 00:17:02.317 }, 00:17:02.317 { 00:17:02.317 "name": "BaseBdev2", 00:17:02.317 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:02.317 "is_configured": true, 00:17:02.317 "data_offset": 256, 00:17:02.317 "data_size": 7936 00:17:02.317 } 00:17:02.317 ] 00:17:02.317 }' 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:02.317 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:02.318 "name": "raid_bdev1", 00:17:02.318 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:02.318 "strip_size_kb": 0, 00:17:02.318 "state": "online", 00:17:02.318 "raid_level": "raid1", 00:17:02.318 "superblock": true, 00:17:02.318 "num_base_bdevs": 2, 00:17:02.318 "num_base_bdevs_discovered": 2, 00:17:02.318 "num_base_bdevs_operational": 2, 00:17:02.318 "base_bdevs_list": [ 00:17:02.318 { 00:17:02.318 "name": "spare", 00:17:02.318 "uuid": "d412ac6c-c84c-55ed-a564-acfe470c4a64", 00:17:02.318 "is_configured": true, 00:17:02.318 "data_offset": 256, 00:17:02.318 "data_size": 7936 00:17:02.318 }, 00:17:02.318 { 00:17:02.318 "name": "BaseBdev2", 00:17:02.318 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:02.318 "is_configured": true, 00:17:02.318 "data_offset": 256, 00:17:02.318 "data_size": 7936 00:17:02.318 } 00:17:02.318 ] 00:17:02.318 }' 00:17:02.318 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:02.318 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:02.576 [2024-11-20 16:05:00.745104] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:02.576 [2024-11-20 16:05:00.745132] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:02.576 [2024-11-20 16:05:00.745194] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:02.576 [2024-11-20 16:05:00.745252] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:02.576 [2024-11-20 16:05:00.745260] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # jq length 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:02.576 16:05:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:17:02.835 /dev/nbd0 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:02.835 1+0 records in 00:17:02.835 1+0 records out 00:17:02.835 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000202352 s, 20.2 MB/s 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:02.835 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:17:03.093 /dev/nbd1 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:03.093 1+0 records in 00:17:03.093 1+0 records out 00:17:03.093 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000219962 s, 18.6 MB/s 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:03.093 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:17:03.351 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:17:03.351 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:17:03.351 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:03.351 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:03.351 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:17:03.351 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:03.351 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:17:03.351 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:03.351 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:03.351 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:03.351 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:03.351 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:03.351 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:03.351 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:17:03.351 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:17:03.351 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:03.351 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:17:03.608 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:17:03.608 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:17:03.608 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:17:03.608 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:03.608 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:03.608 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:17:03.608 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:17:03.608 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:17:03.608 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:17:03.608 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:17:03.608 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:03.608 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:03.608 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:03.608 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:03.608 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:03.608 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:03.608 [2024-11-20 16:05:01.845899] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:03.608 [2024-11-20 16:05:01.845950] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:03.608 [2024-11-20 16:05:01.845972] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:17:03.608 [2024-11-20 16:05:01.845981] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:03.608 [2024-11-20 16:05:01.847856] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:03.608 [2024-11-20 16:05:01.847983] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:03.608 [2024-11-20 16:05:01.848076] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:17:03.609 [2024-11-20 16:05:01.848122] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:03.609 [2024-11-20 16:05:01.848238] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:03.609 spare 00:17:03.609 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:03.609 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:17:03.609 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:03.609 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:03.866 [2024-11-20 16:05:01.948319] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:17:03.866 [2024-11-20 16:05:01.948362] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:17:03.866 [2024-11-20 16:05:01.948638] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1b50 00:17:03.866 [2024-11-20 16:05:01.948818] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:17:03.866 [2024-11-20 16:05:01.948826] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:17:03.866 [2024-11-20 16:05:01.948971] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:03.866 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:03.866 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:03.866 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:03.866 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:03.866 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:03.866 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:03.866 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:03.866 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:03.866 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:03.866 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:03.866 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:03.866 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:03.866 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:03.866 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:03.866 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:03.866 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:03.866 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:03.866 "name": "raid_bdev1", 00:17:03.866 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:03.866 "strip_size_kb": 0, 00:17:03.866 "state": "online", 00:17:03.866 "raid_level": "raid1", 00:17:03.866 "superblock": true, 00:17:03.866 "num_base_bdevs": 2, 00:17:03.866 "num_base_bdevs_discovered": 2, 00:17:03.866 "num_base_bdevs_operational": 2, 00:17:03.866 "base_bdevs_list": [ 00:17:03.866 { 00:17:03.866 "name": "spare", 00:17:03.866 "uuid": "d412ac6c-c84c-55ed-a564-acfe470c4a64", 00:17:03.866 "is_configured": true, 00:17:03.866 "data_offset": 256, 00:17:03.866 "data_size": 7936 00:17:03.866 }, 00:17:03.866 { 00:17:03.866 "name": "BaseBdev2", 00:17:03.866 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:03.866 "is_configured": true, 00:17:03.866 "data_offset": 256, 00:17:03.866 "data_size": 7936 00:17:03.866 } 00:17:03.866 ] 00:17:03.867 }' 00:17:03.867 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:03.867 16:05:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:04.124 "name": "raid_bdev1", 00:17:04.124 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:04.124 "strip_size_kb": 0, 00:17:04.124 "state": "online", 00:17:04.124 "raid_level": "raid1", 00:17:04.124 "superblock": true, 00:17:04.124 "num_base_bdevs": 2, 00:17:04.124 "num_base_bdevs_discovered": 2, 00:17:04.124 "num_base_bdevs_operational": 2, 00:17:04.124 "base_bdevs_list": [ 00:17:04.124 { 00:17:04.124 "name": "spare", 00:17:04.124 "uuid": "d412ac6c-c84c-55ed-a564-acfe470c4a64", 00:17:04.124 "is_configured": true, 00:17:04.124 "data_offset": 256, 00:17:04.124 "data_size": 7936 00:17:04.124 }, 00:17:04.124 { 00:17:04.124 "name": "BaseBdev2", 00:17:04.124 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:04.124 "is_configured": true, 00:17:04.124 "data_offset": 256, 00:17:04.124 "data_size": 7936 00:17:04.124 } 00:17:04.124 ] 00:17:04.124 }' 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:04.124 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:04.382 [2024-11-20 16:05:02.386035] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:04.382 "name": "raid_bdev1", 00:17:04.382 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:04.382 "strip_size_kb": 0, 00:17:04.382 "state": "online", 00:17:04.382 "raid_level": "raid1", 00:17:04.382 "superblock": true, 00:17:04.382 "num_base_bdevs": 2, 00:17:04.382 "num_base_bdevs_discovered": 1, 00:17:04.382 "num_base_bdevs_operational": 1, 00:17:04.382 "base_bdevs_list": [ 00:17:04.382 { 00:17:04.382 "name": null, 00:17:04.382 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:04.382 "is_configured": false, 00:17:04.382 "data_offset": 0, 00:17:04.382 "data_size": 7936 00:17:04.382 }, 00:17:04.382 { 00:17:04.382 "name": "BaseBdev2", 00:17:04.382 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:04.382 "is_configured": true, 00:17:04.382 "data_offset": 256, 00:17:04.382 "data_size": 7936 00:17:04.382 } 00:17:04.382 ] 00:17:04.382 }' 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:04.382 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:04.639 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:04.639 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:04.639 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:04.639 [2024-11-20 16:05:02.718116] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:04.639 [2024-11-20 16:05:02.718280] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:17:04.639 [2024-11-20 16:05:02.718295] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:17:04.639 [2024-11-20 16:05:02.718329] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:04.639 [2024-11-20 16:05:02.727455] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1c20 00:17:04.639 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:04.639 16:05:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@757 -- # sleep 1 00:17:04.639 [2024-11-20 16:05:02.729137] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:05.571 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:05.571 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:05.571 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:05.571 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:05.571 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:05.571 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:05.571 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:05.571 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:05.571 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:05.571 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:05.571 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:05.571 "name": "raid_bdev1", 00:17:05.571 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:05.571 "strip_size_kb": 0, 00:17:05.571 "state": "online", 00:17:05.571 "raid_level": "raid1", 00:17:05.571 "superblock": true, 00:17:05.571 "num_base_bdevs": 2, 00:17:05.571 "num_base_bdevs_discovered": 2, 00:17:05.571 "num_base_bdevs_operational": 2, 00:17:05.571 "process": { 00:17:05.571 "type": "rebuild", 00:17:05.571 "target": "spare", 00:17:05.571 "progress": { 00:17:05.571 "blocks": 2560, 00:17:05.572 "percent": 32 00:17:05.572 } 00:17:05.572 }, 00:17:05.572 "base_bdevs_list": [ 00:17:05.572 { 00:17:05.572 "name": "spare", 00:17:05.572 "uuid": "d412ac6c-c84c-55ed-a564-acfe470c4a64", 00:17:05.572 "is_configured": true, 00:17:05.572 "data_offset": 256, 00:17:05.572 "data_size": 7936 00:17:05.572 }, 00:17:05.572 { 00:17:05.572 "name": "BaseBdev2", 00:17:05.572 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:05.572 "is_configured": true, 00:17:05.572 "data_offset": 256, 00:17:05.572 "data_size": 7936 00:17:05.572 } 00:17:05.572 ] 00:17:05.572 }' 00:17:05.572 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:05.572 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:05.572 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:05.830 [2024-11-20 16:05:03.835447] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:05.830 [2024-11-20 16:05:03.934637] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:05.830 [2024-11-20 16:05:03.934886] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:05.830 [2024-11-20 16:05:03.934903] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:05.830 [2024-11-20 16:05:03.934913] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:05.830 "name": "raid_bdev1", 00:17:05.830 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:05.830 "strip_size_kb": 0, 00:17:05.830 "state": "online", 00:17:05.830 "raid_level": "raid1", 00:17:05.830 "superblock": true, 00:17:05.830 "num_base_bdevs": 2, 00:17:05.830 "num_base_bdevs_discovered": 1, 00:17:05.830 "num_base_bdevs_operational": 1, 00:17:05.830 "base_bdevs_list": [ 00:17:05.830 { 00:17:05.830 "name": null, 00:17:05.830 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:05.830 "is_configured": false, 00:17:05.830 "data_offset": 0, 00:17:05.830 "data_size": 7936 00:17:05.830 }, 00:17:05.830 { 00:17:05.830 "name": "BaseBdev2", 00:17:05.830 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:05.830 "is_configured": true, 00:17:05.830 "data_offset": 256, 00:17:05.830 "data_size": 7936 00:17:05.830 } 00:17:05.830 ] 00:17:05.830 }' 00:17:05.830 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:05.831 16:05:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:06.088 16:05:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:06.088 16:05:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:06.088 16:05:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:06.088 [2024-11-20 16:05:04.245764] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:06.088 [2024-11-20 16:05:04.245922] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:06.088 [2024-11-20 16:05:04.245943] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:17:06.088 [2024-11-20 16:05:04.245953] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:06.088 [2024-11-20 16:05:04.246330] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:06.088 [2024-11-20 16:05:04.246343] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:06.088 [2024-11-20 16:05:04.246417] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:17:06.088 [2024-11-20 16:05:04.246428] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:17:06.088 [2024-11-20 16:05:04.246436] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:17:06.088 [2024-11-20 16:05:04.246456] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:06.088 [2024-11-20 16:05:04.255615] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1cf0 00:17:06.088 spare 00:17:06.088 16:05:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:06.088 16:05:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@764 -- # sleep 1 00:17:06.088 [2024-11-20 16:05:04.257211] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:07.021 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:07.021 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:07.021 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:07.021 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:07.021 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:07.021 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:07.021 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:07.021 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:07.021 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:07.279 "name": "raid_bdev1", 00:17:07.279 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:07.279 "strip_size_kb": 0, 00:17:07.279 "state": "online", 00:17:07.279 "raid_level": "raid1", 00:17:07.279 "superblock": true, 00:17:07.279 "num_base_bdevs": 2, 00:17:07.279 "num_base_bdevs_discovered": 2, 00:17:07.279 "num_base_bdevs_operational": 2, 00:17:07.279 "process": { 00:17:07.279 "type": "rebuild", 00:17:07.279 "target": "spare", 00:17:07.279 "progress": { 00:17:07.279 "blocks": 2560, 00:17:07.279 "percent": 32 00:17:07.279 } 00:17:07.279 }, 00:17:07.279 "base_bdevs_list": [ 00:17:07.279 { 00:17:07.279 "name": "spare", 00:17:07.279 "uuid": "d412ac6c-c84c-55ed-a564-acfe470c4a64", 00:17:07.279 "is_configured": true, 00:17:07.279 "data_offset": 256, 00:17:07.279 "data_size": 7936 00:17:07.279 }, 00:17:07.279 { 00:17:07.279 "name": "BaseBdev2", 00:17:07.279 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:07.279 "is_configured": true, 00:17:07.279 "data_offset": 256, 00:17:07.279 "data_size": 7936 00:17:07.279 } 00:17:07.279 ] 00:17:07.279 }' 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:07.279 [2024-11-20 16:05:05.363499] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:07.279 [2024-11-20 16:05:05.462639] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:07.279 [2024-11-20 16:05:05.462824] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:07.279 [2024-11-20 16:05:05.462887] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:07.279 [2024-11-20 16:05:05.462909] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:07.279 "name": "raid_bdev1", 00:17:07.279 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:07.279 "strip_size_kb": 0, 00:17:07.279 "state": "online", 00:17:07.279 "raid_level": "raid1", 00:17:07.279 "superblock": true, 00:17:07.279 "num_base_bdevs": 2, 00:17:07.279 "num_base_bdevs_discovered": 1, 00:17:07.279 "num_base_bdevs_operational": 1, 00:17:07.279 "base_bdevs_list": [ 00:17:07.279 { 00:17:07.279 "name": null, 00:17:07.279 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:07.279 "is_configured": false, 00:17:07.279 "data_offset": 0, 00:17:07.279 "data_size": 7936 00:17:07.279 }, 00:17:07.279 { 00:17:07.279 "name": "BaseBdev2", 00:17:07.279 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:07.279 "is_configured": true, 00:17:07.279 "data_offset": 256, 00:17:07.279 "data_size": 7936 00:17:07.279 } 00:17:07.279 ] 00:17:07.279 }' 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:07.279 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:07.847 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:07.847 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:07.847 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:07.847 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:07.847 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:07.847 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:07.847 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:07.847 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:07.847 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:07.847 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:07.847 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:07.847 "name": "raid_bdev1", 00:17:07.847 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:07.847 "strip_size_kb": 0, 00:17:07.847 "state": "online", 00:17:07.847 "raid_level": "raid1", 00:17:07.847 "superblock": true, 00:17:07.847 "num_base_bdevs": 2, 00:17:07.847 "num_base_bdevs_discovered": 1, 00:17:07.847 "num_base_bdevs_operational": 1, 00:17:07.847 "base_bdevs_list": [ 00:17:07.847 { 00:17:07.847 "name": null, 00:17:07.847 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:07.847 "is_configured": false, 00:17:07.847 "data_offset": 0, 00:17:07.847 "data_size": 7936 00:17:07.847 }, 00:17:07.847 { 00:17:07.847 "name": "BaseBdev2", 00:17:07.847 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:07.847 "is_configured": true, 00:17:07.847 "data_offset": 256, 00:17:07.847 "data_size": 7936 00:17:07.847 } 00:17:07.847 ] 00:17:07.847 }' 00:17:07.847 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:07.847 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:07.848 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:07.848 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:07.848 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:17:07.848 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:07.848 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:07.848 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:07.848 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:17:07.848 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:07.848 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:07.848 [2024-11-20 16:05:05.913565] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:17:07.848 [2024-11-20 16:05:05.913717] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:07.848 [2024-11-20 16:05:05.913746] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:17:07.848 [2024-11-20 16:05:05.913754] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:07.848 [2024-11-20 16:05:05.914112] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:07.848 [2024-11-20 16:05:05.914124] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:17:07.848 [2024-11-20 16:05:05.914195] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:17:07.848 [2024-11-20 16:05:05.914207] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:17:07.848 [2024-11-20 16:05:05.914215] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:17:07.848 [2024-11-20 16:05:05.914222] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:17:07.848 BaseBdev1 00:17:07.848 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:07.848 16:05:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@775 -- # sleep 1 00:17:08.840 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:08.841 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:08.841 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:08.841 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:08.841 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:08.841 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:08.841 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:08.841 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:08.841 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:08.841 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:08.841 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:08.841 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:08.841 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:08.841 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:08.841 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:08.841 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:08.841 "name": "raid_bdev1", 00:17:08.841 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:08.841 "strip_size_kb": 0, 00:17:08.841 "state": "online", 00:17:08.841 "raid_level": "raid1", 00:17:08.841 "superblock": true, 00:17:08.841 "num_base_bdevs": 2, 00:17:08.841 "num_base_bdevs_discovered": 1, 00:17:08.841 "num_base_bdevs_operational": 1, 00:17:08.841 "base_bdevs_list": [ 00:17:08.841 { 00:17:08.841 "name": null, 00:17:08.841 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:08.841 "is_configured": false, 00:17:08.841 "data_offset": 0, 00:17:08.841 "data_size": 7936 00:17:08.841 }, 00:17:08.841 { 00:17:08.841 "name": "BaseBdev2", 00:17:08.841 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:08.841 "is_configured": true, 00:17:08.841 "data_offset": 256, 00:17:08.841 "data_size": 7936 00:17:08.841 } 00:17:08.841 ] 00:17:08.841 }' 00:17:08.841 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:08.841 16:05:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:09.098 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:09.098 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:09.098 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:09.098 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:09.098 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:09.099 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:09.099 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:09.099 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:09.099 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:09.099 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:09.099 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:09.099 "name": "raid_bdev1", 00:17:09.099 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:09.099 "strip_size_kb": 0, 00:17:09.099 "state": "online", 00:17:09.099 "raid_level": "raid1", 00:17:09.099 "superblock": true, 00:17:09.099 "num_base_bdevs": 2, 00:17:09.099 "num_base_bdevs_discovered": 1, 00:17:09.099 "num_base_bdevs_operational": 1, 00:17:09.099 "base_bdevs_list": [ 00:17:09.099 { 00:17:09.099 "name": null, 00:17:09.099 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:09.099 "is_configured": false, 00:17:09.099 "data_offset": 0, 00:17:09.099 "data_size": 7936 00:17:09.099 }, 00:17:09.099 { 00:17:09.099 "name": "BaseBdev2", 00:17:09.099 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:09.099 "is_configured": true, 00:17:09.099 "data_offset": 256, 00:17:09.099 "data_size": 7936 00:17:09.099 } 00:17:09.099 ] 00:17:09.099 }' 00:17:09.099 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:09.099 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:09.099 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:09.361 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:09.361 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:09.361 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@652 -- # local es=0 00:17:09.361 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:09.361 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:17:09.361 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:09.361 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:17:09.361 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:09.361 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:09.361 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:09.361 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:09.361 [2024-11-20 16:05:07.361889] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:09.361 [2024-11-20 16:05:07.362014] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:17:09.361 [2024-11-20 16:05:07.362028] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:17:09.361 request: 00:17:09.361 { 00:17:09.361 "base_bdev": "BaseBdev1", 00:17:09.361 "raid_bdev": "raid_bdev1", 00:17:09.361 "method": "bdev_raid_add_base_bdev", 00:17:09.361 "req_id": 1 00:17:09.361 } 00:17:09.361 Got JSON-RPC error response 00:17:09.361 response: 00:17:09.361 { 00:17:09.361 "code": -22, 00:17:09.361 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:17:09.361 } 00:17:09.361 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:17:09.361 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@655 -- # es=1 00:17:09.361 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:17:09.361 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:17:09.361 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:17:09.361 16:05:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@779 -- # sleep 1 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:10.294 "name": "raid_bdev1", 00:17:10.294 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:10.294 "strip_size_kb": 0, 00:17:10.294 "state": "online", 00:17:10.294 "raid_level": "raid1", 00:17:10.294 "superblock": true, 00:17:10.294 "num_base_bdevs": 2, 00:17:10.294 "num_base_bdevs_discovered": 1, 00:17:10.294 "num_base_bdevs_operational": 1, 00:17:10.294 "base_bdevs_list": [ 00:17:10.294 { 00:17:10.294 "name": null, 00:17:10.294 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:10.294 "is_configured": false, 00:17:10.294 "data_offset": 0, 00:17:10.294 "data_size": 7936 00:17:10.294 }, 00:17:10.294 { 00:17:10.294 "name": "BaseBdev2", 00:17:10.294 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:10.294 "is_configured": true, 00:17:10.294 "data_offset": 256, 00:17:10.294 "data_size": 7936 00:17:10.294 } 00:17:10.294 ] 00:17:10.294 }' 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:10.294 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:10.552 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:10.552 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:10.552 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:10.552 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:10.552 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:10.552 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:10.552 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:10.552 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:10.552 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:10.552 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:10.552 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:10.552 "name": "raid_bdev1", 00:17:10.552 "uuid": "4c6b43f8-c06f-4c92-aef5-130376301dd0", 00:17:10.552 "strip_size_kb": 0, 00:17:10.552 "state": "online", 00:17:10.552 "raid_level": "raid1", 00:17:10.552 "superblock": true, 00:17:10.552 "num_base_bdevs": 2, 00:17:10.552 "num_base_bdevs_discovered": 1, 00:17:10.552 "num_base_bdevs_operational": 1, 00:17:10.552 "base_bdevs_list": [ 00:17:10.552 { 00:17:10.552 "name": null, 00:17:10.552 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:10.552 "is_configured": false, 00:17:10.552 "data_offset": 0, 00:17:10.552 "data_size": 7936 00:17:10.552 }, 00:17:10.552 { 00:17:10.552 "name": "BaseBdev2", 00:17:10.552 "uuid": "87b9a405-9411-501d-a584-6a1454fa7ea4", 00:17:10.552 "is_configured": true, 00:17:10.552 "data_offset": 256, 00:17:10.552 "data_size": 7936 00:17:10.552 } 00:17:10.552 ] 00:17:10.552 }' 00:17:10.552 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:10.552 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:10.552 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:10.810 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:10.810 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@784 -- # killprocess 84068 00:17:10.810 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@954 -- # '[' -z 84068 ']' 00:17:10.810 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@958 -- # kill -0 84068 00:17:10.810 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@959 -- # uname 00:17:10.810 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:10.810 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 84068 00:17:10.810 killing process with pid 84068 00:17:10.810 Received shutdown signal, test time was about 60.000000 seconds 00:17:10.810 00:17:10.810 Latency(us) 00:17:10.810 [2024-11-20T16:05:09.061Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:10.810 [2024-11-20T16:05:09.061Z] =================================================================================================================== 00:17:10.810 [2024-11-20T16:05:09.061Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:17:10.810 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:10.810 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:10.810 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 84068' 00:17:10.810 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@973 -- # kill 84068 00:17:10.810 [2024-11-20 16:05:08.831434] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:10.810 16:05:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@978 -- # wait 84068 00:17:10.810 [2024-11-20 16:05:08.831528] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:10.810 [2024-11-20 16:05:08.831566] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:10.810 [2024-11-20 16:05:08.831575] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:17:10.810 [2024-11-20 16:05:08.980504] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:11.379 16:05:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@786 -- # return 0 00:17:11.379 ************************************ 00:17:11.379 END TEST raid_rebuild_test_sb_4k 00:17:11.379 ************************************ 00:17:11.379 00:17:11.379 real 0m16.848s 00:17:11.379 user 0m21.443s 00:17:11.379 sys 0m1.771s 00:17:11.379 16:05:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:11.379 16:05:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:11.380 16:05:09 bdev_raid -- bdev/bdev_raid.sh@1003 -- # base_malloc_params='-m 32' 00:17:11.380 16:05:09 bdev_raid -- bdev/bdev_raid.sh@1004 -- # run_test raid_state_function_test_sb_md_separate raid_state_function_test raid1 2 true 00:17:11.380 16:05:09 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:17:11.380 16:05:09 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:11.380 16:05:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:11.380 ************************************ 00:17:11.380 START TEST raid_state_function_test_sb_md_separate 00:17:11.380 ************************************ 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # local strip_size 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:17:11.380 Process raid pid: 84725 00:17:11.380 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@229 -- # raid_pid=84725 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 84725' 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@231 -- # waitforlisten 84725 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@835 -- # '[' -z 84725 ']' 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:11.380 16:05:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:11.639 [2024-11-20 16:05:09.670187] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:17:11.639 [2024-11-20 16:05:09.670465] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:17:11.639 [2024-11-20 16:05:09.827443] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:11.897 [2024-11-20 16:05:09.912780] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:11.897 [2024-11-20 16:05:10.024049] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:11.897 [2024-11-20 16:05:10.024208] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@868 -- # return 0 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:12.495 [2024-11-20 16:05:10.520445] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:17:12.495 [2024-11-20 16:05:10.520488] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:17:12.495 [2024-11-20 16:05:10.520497] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:17:12.495 [2024-11-20 16:05:10.520505] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:12.495 "name": "Existed_Raid", 00:17:12.495 "uuid": "831f8c0f-5b81-4aca-ab21-d4ee0eaecdcd", 00:17:12.495 "strip_size_kb": 0, 00:17:12.495 "state": "configuring", 00:17:12.495 "raid_level": "raid1", 00:17:12.495 "superblock": true, 00:17:12.495 "num_base_bdevs": 2, 00:17:12.495 "num_base_bdevs_discovered": 0, 00:17:12.495 "num_base_bdevs_operational": 2, 00:17:12.495 "base_bdevs_list": [ 00:17:12.495 { 00:17:12.495 "name": "BaseBdev1", 00:17:12.495 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:12.495 "is_configured": false, 00:17:12.495 "data_offset": 0, 00:17:12.495 "data_size": 0 00:17:12.495 }, 00:17:12.495 { 00:17:12.495 "name": "BaseBdev2", 00:17:12.495 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:12.495 "is_configured": false, 00:17:12.495 "data_offset": 0, 00:17:12.495 "data_size": 0 00:17:12.495 } 00:17:12.495 ] 00:17:12.495 }' 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:12.495 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:12.753 [2024-11-20 16:05:10.852472] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:17:12.753 [2024-11-20 16:05:10.852500] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:12.753 [2024-11-20 16:05:10.860464] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:17:12.753 [2024-11-20 16:05:10.860499] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:17:12.753 [2024-11-20 16:05:10.860506] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:17:12.753 [2024-11-20 16:05:10.860516] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:12.753 [2024-11-20 16:05:10.888985] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:12.753 BaseBdev1 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@905 -- # local i 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:12.753 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:12.753 [ 00:17:12.753 { 00:17:12.753 "name": "BaseBdev1", 00:17:12.753 "aliases": [ 00:17:12.753 "49276686-d1d5-41cf-b9b8-5b7463e0743f" 00:17:12.753 ], 00:17:12.753 "product_name": "Malloc disk", 00:17:12.753 "block_size": 4096, 00:17:12.753 "num_blocks": 8192, 00:17:12.753 "uuid": "49276686-d1d5-41cf-b9b8-5b7463e0743f", 00:17:12.753 "md_size": 32, 00:17:12.753 "md_interleave": false, 00:17:12.753 "dif_type": 0, 00:17:12.753 "assigned_rate_limits": { 00:17:12.753 "rw_ios_per_sec": 0, 00:17:12.753 "rw_mbytes_per_sec": 0, 00:17:12.753 "r_mbytes_per_sec": 0, 00:17:12.753 "w_mbytes_per_sec": 0 00:17:12.753 }, 00:17:12.753 "claimed": true, 00:17:12.754 "claim_type": "exclusive_write", 00:17:12.754 "zoned": false, 00:17:12.754 "supported_io_types": { 00:17:12.754 "read": true, 00:17:12.754 "write": true, 00:17:12.754 "unmap": true, 00:17:12.754 "flush": true, 00:17:12.754 "reset": true, 00:17:12.754 "nvme_admin": false, 00:17:12.754 "nvme_io": false, 00:17:12.754 "nvme_io_md": false, 00:17:12.754 "write_zeroes": true, 00:17:12.754 "zcopy": true, 00:17:12.754 "get_zone_info": false, 00:17:12.754 "zone_management": false, 00:17:12.754 "zone_append": false, 00:17:12.754 "compare": false, 00:17:12.754 "compare_and_write": false, 00:17:12.754 "abort": true, 00:17:12.754 "seek_hole": false, 00:17:12.754 "seek_data": false, 00:17:12.754 "copy": true, 00:17:12.754 "nvme_iov_md": false 00:17:12.754 }, 00:17:12.754 "memory_domains": [ 00:17:12.754 { 00:17:12.754 "dma_device_id": "system", 00:17:12.754 "dma_device_type": 1 00:17:12.754 }, 00:17:12.754 { 00:17:12.754 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:12.754 "dma_device_type": 2 00:17:12.754 } 00:17:12.754 ], 00:17:12.754 "driver_specific": {} 00:17:12.754 } 00:17:12.754 ] 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@911 -- # return 0 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:12.754 "name": "Existed_Raid", 00:17:12.754 "uuid": "107f61a5-e667-445f-b2b7-1be6efa17796", 00:17:12.754 "strip_size_kb": 0, 00:17:12.754 "state": "configuring", 00:17:12.754 "raid_level": "raid1", 00:17:12.754 "superblock": true, 00:17:12.754 "num_base_bdevs": 2, 00:17:12.754 "num_base_bdevs_discovered": 1, 00:17:12.754 "num_base_bdevs_operational": 2, 00:17:12.754 "base_bdevs_list": [ 00:17:12.754 { 00:17:12.754 "name": "BaseBdev1", 00:17:12.754 "uuid": "49276686-d1d5-41cf-b9b8-5b7463e0743f", 00:17:12.754 "is_configured": true, 00:17:12.754 "data_offset": 256, 00:17:12.754 "data_size": 7936 00:17:12.754 }, 00:17:12.754 { 00:17:12.754 "name": "BaseBdev2", 00:17:12.754 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:12.754 "is_configured": false, 00:17:12.754 "data_offset": 0, 00:17:12.754 "data_size": 0 00:17:12.754 } 00:17:12.754 ] 00:17:12.754 }' 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:12.754 16:05:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:13.012 [2024-11-20 16:05:11.229119] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:17:13.012 [2024-11-20 16:05:11.229159] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:13.012 [2024-11-20 16:05:11.237145] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:13.012 [2024-11-20 16:05:11.238688] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:17:13.012 [2024-11-20 16:05:11.238817] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:13.012 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:13.270 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:13.270 "name": "Existed_Raid", 00:17:13.270 "uuid": "b442bd02-f936-46b1-8bc7-ce04417b09fc", 00:17:13.270 "strip_size_kb": 0, 00:17:13.270 "state": "configuring", 00:17:13.270 "raid_level": "raid1", 00:17:13.270 "superblock": true, 00:17:13.270 "num_base_bdevs": 2, 00:17:13.270 "num_base_bdevs_discovered": 1, 00:17:13.270 "num_base_bdevs_operational": 2, 00:17:13.270 "base_bdevs_list": [ 00:17:13.270 { 00:17:13.270 "name": "BaseBdev1", 00:17:13.270 "uuid": "49276686-d1d5-41cf-b9b8-5b7463e0743f", 00:17:13.270 "is_configured": true, 00:17:13.270 "data_offset": 256, 00:17:13.270 "data_size": 7936 00:17:13.270 }, 00:17:13.270 { 00:17:13.270 "name": "BaseBdev2", 00:17:13.270 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:13.270 "is_configured": false, 00:17:13.270 "data_offset": 0, 00:17:13.270 "data_size": 0 00:17:13.270 } 00:17:13.270 ] 00:17:13.270 }' 00:17:13.270 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:13.270 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:13.529 [2024-11-20 16:05:11.600388] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:13.529 [2024-11-20 16:05:11.600605] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:17:13.529 [2024-11-20 16:05:11.600622] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:17:13.529 [2024-11-20 16:05:11.600719] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:17:13.529 [2024-11-20 16:05:11.600835] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:17:13.529 [2024-11-20 16:05:11.600845] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:17:13.529 BaseBdev2 00:17:13.529 [2024-11-20 16:05:11.600928] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@905 -- # local i 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:13.529 [ 00:17:13.529 { 00:17:13.529 "name": "BaseBdev2", 00:17:13.529 "aliases": [ 00:17:13.529 "416cc077-5331-42d4-a577-6cc8cc28f22f" 00:17:13.529 ], 00:17:13.529 "product_name": "Malloc disk", 00:17:13.529 "block_size": 4096, 00:17:13.529 "num_blocks": 8192, 00:17:13.529 "uuid": "416cc077-5331-42d4-a577-6cc8cc28f22f", 00:17:13.529 "md_size": 32, 00:17:13.529 "md_interleave": false, 00:17:13.529 "dif_type": 0, 00:17:13.529 "assigned_rate_limits": { 00:17:13.529 "rw_ios_per_sec": 0, 00:17:13.529 "rw_mbytes_per_sec": 0, 00:17:13.529 "r_mbytes_per_sec": 0, 00:17:13.529 "w_mbytes_per_sec": 0 00:17:13.529 }, 00:17:13.529 "claimed": true, 00:17:13.529 "claim_type": "exclusive_write", 00:17:13.529 "zoned": false, 00:17:13.529 "supported_io_types": { 00:17:13.529 "read": true, 00:17:13.529 "write": true, 00:17:13.529 "unmap": true, 00:17:13.529 "flush": true, 00:17:13.529 "reset": true, 00:17:13.529 "nvme_admin": false, 00:17:13.529 "nvme_io": false, 00:17:13.529 "nvme_io_md": false, 00:17:13.529 "write_zeroes": true, 00:17:13.529 "zcopy": true, 00:17:13.529 "get_zone_info": false, 00:17:13.529 "zone_management": false, 00:17:13.529 "zone_append": false, 00:17:13.529 "compare": false, 00:17:13.529 "compare_and_write": false, 00:17:13.529 "abort": true, 00:17:13.529 "seek_hole": false, 00:17:13.529 "seek_data": false, 00:17:13.529 "copy": true, 00:17:13.529 "nvme_iov_md": false 00:17:13.529 }, 00:17:13.529 "memory_domains": [ 00:17:13.529 { 00:17:13.529 "dma_device_id": "system", 00:17:13.529 "dma_device_type": 1 00:17:13.529 }, 00:17:13.529 { 00:17:13.529 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:13.529 "dma_device_type": 2 00:17:13.529 } 00:17:13.529 ], 00:17:13.529 "driver_specific": {} 00:17:13.529 } 00:17:13.529 ] 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@911 -- # return 0 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:13.529 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:13.529 "name": "Existed_Raid", 00:17:13.529 "uuid": "b442bd02-f936-46b1-8bc7-ce04417b09fc", 00:17:13.529 "strip_size_kb": 0, 00:17:13.529 "state": "online", 00:17:13.529 "raid_level": "raid1", 00:17:13.529 "superblock": true, 00:17:13.529 "num_base_bdevs": 2, 00:17:13.529 "num_base_bdevs_discovered": 2, 00:17:13.529 "num_base_bdevs_operational": 2, 00:17:13.529 "base_bdevs_list": [ 00:17:13.529 { 00:17:13.529 "name": "BaseBdev1", 00:17:13.529 "uuid": "49276686-d1d5-41cf-b9b8-5b7463e0743f", 00:17:13.529 "is_configured": true, 00:17:13.529 "data_offset": 256, 00:17:13.530 "data_size": 7936 00:17:13.530 }, 00:17:13.530 { 00:17:13.530 "name": "BaseBdev2", 00:17:13.530 "uuid": "416cc077-5331-42d4-a577-6cc8cc28f22f", 00:17:13.530 "is_configured": true, 00:17:13.530 "data_offset": 256, 00:17:13.530 "data_size": 7936 00:17:13.530 } 00:17:13.530 ] 00:17:13.530 }' 00:17:13.530 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:13.530 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:13.789 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:17:13.789 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:17:13.789 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:17:13.789 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:17:13.789 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:17:13.789 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:17:13.789 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:17:13.789 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:13.789 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:13.789 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:17:13.789 [2024-11-20 16:05:11.944871] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:13.789 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:13.789 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:17:13.789 "name": "Existed_Raid", 00:17:13.789 "aliases": [ 00:17:13.789 "b442bd02-f936-46b1-8bc7-ce04417b09fc" 00:17:13.789 ], 00:17:13.789 "product_name": "Raid Volume", 00:17:13.789 "block_size": 4096, 00:17:13.789 "num_blocks": 7936, 00:17:13.789 "uuid": "b442bd02-f936-46b1-8bc7-ce04417b09fc", 00:17:13.789 "md_size": 32, 00:17:13.789 "md_interleave": false, 00:17:13.789 "dif_type": 0, 00:17:13.789 "assigned_rate_limits": { 00:17:13.789 "rw_ios_per_sec": 0, 00:17:13.789 "rw_mbytes_per_sec": 0, 00:17:13.789 "r_mbytes_per_sec": 0, 00:17:13.789 "w_mbytes_per_sec": 0 00:17:13.789 }, 00:17:13.789 "claimed": false, 00:17:13.789 "zoned": false, 00:17:13.789 "supported_io_types": { 00:17:13.789 "read": true, 00:17:13.789 "write": true, 00:17:13.789 "unmap": false, 00:17:13.789 "flush": false, 00:17:13.789 "reset": true, 00:17:13.789 "nvme_admin": false, 00:17:13.789 "nvme_io": false, 00:17:13.789 "nvme_io_md": false, 00:17:13.789 "write_zeroes": true, 00:17:13.789 "zcopy": false, 00:17:13.789 "get_zone_info": false, 00:17:13.789 "zone_management": false, 00:17:13.789 "zone_append": false, 00:17:13.789 "compare": false, 00:17:13.789 "compare_and_write": false, 00:17:13.789 "abort": false, 00:17:13.789 "seek_hole": false, 00:17:13.789 "seek_data": false, 00:17:13.789 "copy": false, 00:17:13.789 "nvme_iov_md": false 00:17:13.789 }, 00:17:13.789 "memory_domains": [ 00:17:13.789 { 00:17:13.789 "dma_device_id": "system", 00:17:13.789 "dma_device_type": 1 00:17:13.789 }, 00:17:13.789 { 00:17:13.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:13.789 "dma_device_type": 2 00:17:13.789 }, 00:17:13.789 { 00:17:13.789 "dma_device_id": "system", 00:17:13.789 "dma_device_type": 1 00:17:13.789 }, 00:17:13.789 { 00:17:13.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:13.789 "dma_device_type": 2 00:17:13.789 } 00:17:13.790 ], 00:17:13.790 "driver_specific": { 00:17:13.790 "raid": { 00:17:13.790 "uuid": "b442bd02-f936-46b1-8bc7-ce04417b09fc", 00:17:13.790 "strip_size_kb": 0, 00:17:13.790 "state": "online", 00:17:13.790 "raid_level": "raid1", 00:17:13.790 "superblock": true, 00:17:13.790 "num_base_bdevs": 2, 00:17:13.790 "num_base_bdevs_discovered": 2, 00:17:13.790 "num_base_bdevs_operational": 2, 00:17:13.790 "base_bdevs_list": [ 00:17:13.790 { 00:17:13.790 "name": "BaseBdev1", 00:17:13.790 "uuid": "49276686-d1d5-41cf-b9b8-5b7463e0743f", 00:17:13.790 "is_configured": true, 00:17:13.790 "data_offset": 256, 00:17:13.790 "data_size": 7936 00:17:13.790 }, 00:17:13.790 { 00:17:13.790 "name": "BaseBdev2", 00:17:13.790 "uuid": "416cc077-5331-42d4-a577-6cc8cc28f22f", 00:17:13.790 "is_configured": true, 00:17:13.790 "data_offset": 256, 00:17:13.790 "data_size": 7936 00:17:13.790 } 00:17:13.790 ] 00:17:13.790 } 00:17:13.790 } 00:17:13.790 }' 00:17:13.790 16:05:11 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:17:13.790 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:17:13.790 BaseBdev2' 00:17:13.790 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:13.790 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:17:13.790 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:14.049 [2024-11-20 16:05:12.104602] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@260 -- # local expected_state 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:14.049 "name": "Existed_Raid", 00:17:14.049 "uuid": "b442bd02-f936-46b1-8bc7-ce04417b09fc", 00:17:14.049 "strip_size_kb": 0, 00:17:14.049 "state": "online", 00:17:14.049 "raid_level": "raid1", 00:17:14.049 "superblock": true, 00:17:14.049 "num_base_bdevs": 2, 00:17:14.049 "num_base_bdevs_discovered": 1, 00:17:14.049 "num_base_bdevs_operational": 1, 00:17:14.049 "base_bdevs_list": [ 00:17:14.049 { 00:17:14.049 "name": null, 00:17:14.049 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:14.049 "is_configured": false, 00:17:14.049 "data_offset": 0, 00:17:14.049 "data_size": 7936 00:17:14.049 }, 00:17:14.049 { 00:17:14.049 "name": "BaseBdev2", 00:17:14.049 "uuid": "416cc077-5331-42d4-a577-6cc8cc28f22f", 00:17:14.049 "is_configured": true, 00:17:14.049 "data_offset": 256, 00:17:14.049 "data_size": 7936 00:17:14.049 } 00:17:14.049 ] 00:17:14.049 }' 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:14.049 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:14.305 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:17:14.305 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:17:14.305 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:14.305 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:14.305 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:14.305 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:17:14.306 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:14.306 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:17:14.306 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:17:14.306 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:17:14.306 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:14.306 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:14.306 [2024-11-20 16:05:12.516652] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:17:14.306 [2024-11-20 16:05:12.516759] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:14.563 [2024-11-20 16:05:12.581721] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:14.563 [2024-11-20 16:05:12.581909] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:14.563 [2024-11-20 16:05:12.581928] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@326 -- # killprocess 84725 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@954 -- # '[' -z 84725 ']' 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@958 -- # kill -0 84725 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@959 -- # uname 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 84725 00:17:14.563 killing process with pid 84725 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 84725' 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@973 -- # kill 84725 00:17:14.563 [2024-11-20 16:05:12.641889] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:14.563 16:05:12 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@978 -- # wait 84725 00:17:14.563 [2024-11-20 16:05:12.652623] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:15.499 16:05:13 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@328 -- # return 0 00:17:15.499 00:17:15.499 real 0m3.776s 00:17:15.499 user 0m5.433s 00:17:15.499 sys 0m0.568s 00:17:15.499 16:05:13 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:15.499 16:05:13 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:15.499 ************************************ 00:17:15.499 END TEST raid_state_function_test_sb_md_separate 00:17:15.499 ************************************ 00:17:15.499 16:05:13 bdev_raid -- bdev/bdev_raid.sh@1005 -- # run_test raid_superblock_test_md_separate raid_superblock_test raid1 2 00:17:15.499 16:05:13 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:17:15.499 16:05:13 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:15.499 16:05:13 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:15.499 ************************************ 00:17:15.499 START TEST raid_superblock_test_md_separate 00:17:15.499 ************************************ 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@399 -- # local strip_size 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@412 -- # raid_pid=84966 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@413 -- # waitforlisten 84966 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@835 -- # '[' -z 84966 ']' 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:15.499 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:15.499 16:05:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:15.499 [2024-11-20 16:05:13.490999] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:17:15.499 [2024-11-20 16:05:13.491253] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid84966 ] 00:17:15.499 [2024-11-20 16:05:13.645133] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:15.760 [2024-11-20 16:05:13.749602] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:15.760 [2024-11-20 16:05:13.888665] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:15.760 [2024-11-20 16:05:13.888712] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:16.332 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:16.332 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@868 -- # return 0 00:17:16.332 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:17:16.332 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:17:16.332 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:17:16.332 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:17:16.332 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:17:16.332 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:17:16.332 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:17:16.332 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:17:16.332 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc1 00:17:16.332 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.332 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.332 malloc1 00:17:16.332 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.332 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:17:16.332 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.332 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.332 [2024-11-20 16:05:14.394707] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:17:16.332 [2024-11-20 16:05:14.394760] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:16.332 [2024-11-20 16:05:14.394781] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:17:16.332 [2024-11-20 16:05:14.394790] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:16.332 [2024-11-20 16:05:14.396714] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:16.333 [2024-11-20 16:05:14.396744] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:17:16.333 pt1 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc2 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.333 malloc2 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.333 [2024-11-20 16:05:14.435125] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:16.333 [2024-11-20 16:05:14.435174] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:16.333 [2024-11-20 16:05:14.435194] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:17:16.333 [2024-11-20 16:05:14.435202] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:16.333 [2024-11-20 16:05:14.437089] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:16.333 [2024-11-20 16:05:14.437119] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:16.333 pt2 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.333 [2024-11-20 16:05:14.443157] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:17:16.333 [2024-11-20 16:05:14.444974] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:16.333 [2024-11-20 16:05:14.445138] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:17:16.333 [2024-11-20 16:05:14.445151] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:17:16.333 [2024-11-20 16:05:14.445224] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:17:16.333 [2024-11-20 16:05:14.445336] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:17:16.333 [2024-11-20 16:05:14.445349] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:17:16.333 [2024-11-20 16:05:14.445441] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:16.333 "name": "raid_bdev1", 00:17:16.333 "uuid": "ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded", 00:17:16.333 "strip_size_kb": 0, 00:17:16.333 "state": "online", 00:17:16.333 "raid_level": "raid1", 00:17:16.333 "superblock": true, 00:17:16.333 "num_base_bdevs": 2, 00:17:16.333 "num_base_bdevs_discovered": 2, 00:17:16.333 "num_base_bdevs_operational": 2, 00:17:16.333 "base_bdevs_list": [ 00:17:16.333 { 00:17:16.333 "name": "pt1", 00:17:16.333 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:16.333 "is_configured": true, 00:17:16.333 "data_offset": 256, 00:17:16.333 "data_size": 7936 00:17:16.333 }, 00:17:16.333 { 00:17:16.333 "name": "pt2", 00:17:16.333 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:16.333 "is_configured": true, 00:17:16.333 "data_offset": 256, 00:17:16.333 "data_size": 7936 00:17:16.333 } 00:17:16.333 ] 00:17:16.333 }' 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:16.333 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.621 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:17:16.621 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:17:16.621 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:17:16.621 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:17:16.621 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:17:16.621 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:17:16.621 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:17:16.621 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:16.621 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.621 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.621 [2024-11-20 16:05:14.835548] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:16.621 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.896 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:17:16.896 "name": "raid_bdev1", 00:17:16.896 "aliases": [ 00:17:16.896 "ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded" 00:17:16.896 ], 00:17:16.896 "product_name": "Raid Volume", 00:17:16.896 "block_size": 4096, 00:17:16.896 "num_blocks": 7936, 00:17:16.896 "uuid": "ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded", 00:17:16.896 "md_size": 32, 00:17:16.896 "md_interleave": false, 00:17:16.896 "dif_type": 0, 00:17:16.896 "assigned_rate_limits": { 00:17:16.896 "rw_ios_per_sec": 0, 00:17:16.896 "rw_mbytes_per_sec": 0, 00:17:16.896 "r_mbytes_per_sec": 0, 00:17:16.896 "w_mbytes_per_sec": 0 00:17:16.896 }, 00:17:16.896 "claimed": false, 00:17:16.896 "zoned": false, 00:17:16.896 "supported_io_types": { 00:17:16.896 "read": true, 00:17:16.897 "write": true, 00:17:16.897 "unmap": false, 00:17:16.897 "flush": false, 00:17:16.897 "reset": true, 00:17:16.897 "nvme_admin": false, 00:17:16.897 "nvme_io": false, 00:17:16.897 "nvme_io_md": false, 00:17:16.897 "write_zeroes": true, 00:17:16.897 "zcopy": false, 00:17:16.897 "get_zone_info": false, 00:17:16.897 "zone_management": false, 00:17:16.897 "zone_append": false, 00:17:16.897 "compare": false, 00:17:16.897 "compare_and_write": false, 00:17:16.897 "abort": false, 00:17:16.897 "seek_hole": false, 00:17:16.897 "seek_data": false, 00:17:16.897 "copy": false, 00:17:16.897 "nvme_iov_md": false 00:17:16.897 }, 00:17:16.897 "memory_domains": [ 00:17:16.897 { 00:17:16.897 "dma_device_id": "system", 00:17:16.897 "dma_device_type": 1 00:17:16.897 }, 00:17:16.897 { 00:17:16.897 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:16.897 "dma_device_type": 2 00:17:16.897 }, 00:17:16.897 { 00:17:16.897 "dma_device_id": "system", 00:17:16.897 "dma_device_type": 1 00:17:16.897 }, 00:17:16.897 { 00:17:16.897 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:16.897 "dma_device_type": 2 00:17:16.897 } 00:17:16.897 ], 00:17:16.897 "driver_specific": { 00:17:16.897 "raid": { 00:17:16.897 "uuid": "ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded", 00:17:16.897 "strip_size_kb": 0, 00:17:16.897 "state": "online", 00:17:16.897 "raid_level": "raid1", 00:17:16.897 "superblock": true, 00:17:16.897 "num_base_bdevs": 2, 00:17:16.897 "num_base_bdevs_discovered": 2, 00:17:16.897 "num_base_bdevs_operational": 2, 00:17:16.897 "base_bdevs_list": [ 00:17:16.897 { 00:17:16.897 "name": "pt1", 00:17:16.897 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:16.897 "is_configured": true, 00:17:16.897 "data_offset": 256, 00:17:16.897 "data_size": 7936 00:17:16.897 }, 00:17:16.897 { 00:17:16.897 "name": "pt2", 00:17:16.897 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:16.897 "is_configured": true, 00:17:16.897 "data_offset": 256, 00:17:16.897 "data_size": 7936 00:17:16.897 } 00:17:16.897 ] 00:17:16.897 } 00:17:16.897 } 00:17:16.897 }' 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:17:16.897 pt2' 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.897 16:05:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.897 [2024-11-20 16:05:14.999539] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@436 -- # '[' -z ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded ']' 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.897 [2024-11-20 16:05:15.031234] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:16.897 [2024-11-20 16:05:15.031257] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:16.897 [2024-11-20 16:05:15.031346] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:16.897 [2024-11-20 16:05:15.031407] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:16.897 [2024-11-20 16:05:15.031420] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@652 -- # local es=0 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.897 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.897 [2024-11-20 16:05:15.135286] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:17:16.897 [2024-11-20 16:05:15.137214] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:17:16.897 [2024-11-20 16:05:15.137282] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:17:16.898 [2024-11-20 16:05:15.137332] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:17:16.898 [2024-11-20 16:05:15.137346] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:16.898 [2024-11-20 16:05:15.137357] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:17:16.898 request: 00:17:16.898 { 00:17:16.898 "name": "raid_bdev1", 00:17:16.898 "raid_level": "raid1", 00:17:16.898 "base_bdevs": [ 00:17:16.898 "malloc1", 00:17:16.898 "malloc2" 00:17:16.898 ], 00:17:16.898 "superblock": false, 00:17:16.898 "method": "bdev_raid_create", 00:17:16.898 "req_id": 1 00:17:16.898 } 00:17:16.898 Got JSON-RPC error response 00:17:16.898 response: 00:17:16.898 { 00:17:16.898 "code": -17, 00:17:16.898 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:17:16.898 } 00:17:16.898 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:17:16.898 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@655 -- # es=1 00:17:16.898 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:17:16.898 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:17:16.898 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:17:17.158 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:17.158 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:17:17.158 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.158 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.158 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.158 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:17:17.158 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:17:17.158 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:17:17.158 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.158 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.158 [2024-11-20 16:05:15.183265] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:17:17.158 [2024-11-20 16:05:15.183420] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:17.158 [2024-11-20 16:05:15.183457] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:17:17.158 [2024-11-20 16:05:15.183549] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:17.158 [2024-11-20 16:05:15.185539] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:17.158 [2024-11-20 16:05:15.185651] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:17:17.158 [2024-11-20 16:05:15.185759] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:17:17.159 [2024-11-20 16:05:15.185833] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:17:17.159 pt1 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:17.159 "name": "raid_bdev1", 00:17:17.159 "uuid": "ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded", 00:17:17.159 "strip_size_kb": 0, 00:17:17.159 "state": "configuring", 00:17:17.159 "raid_level": "raid1", 00:17:17.159 "superblock": true, 00:17:17.159 "num_base_bdevs": 2, 00:17:17.159 "num_base_bdevs_discovered": 1, 00:17:17.159 "num_base_bdevs_operational": 2, 00:17:17.159 "base_bdevs_list": [ 00:17:17.159 { 00:17:17.159 "name": "pt1", 00:17:17.159 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:17.159 "is_configured": true, 00:17:17.159 "data_offset": 256, 00:17:17.159 "data_size": 7936 00:17:17.159 }, 00:17:17.159 { 00:17:17.159 "name": null, 00:17:17.159 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:17.159 "is_configured": false, 00:17:17.159 "data_offset": 256, 00:17:17.159 "data_size": 7936 00:17:17.159 } 00:17:17.159 ] 00:17:17.159 }' 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:17.159 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.420 [2024-11-20 16:05:15.503350] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:17.420 [2024-11-20 16:05:15.503526] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:17.420 [2024-11-20 16:05:15.503551] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:17:17.420 [2024-11-20 16:05:15.503562] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:17.420 [2024-11-20 16:05:15.503774] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:17.420 [2024-11-20 16:05:15.503792] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:17.420 [2024-11-20 16:05:15.503839] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:17:17.420 [2024-11-20 16:05:15.503860] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:17.420 [2024-11-20 16:05:15.503964] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:17:17.420 [2024-11-20 16:05:15.503976] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:17:17.420 [2024-11-20 16:05:15.504051] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:17:17.420 [2024-11-20 16:05:15.504150] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:17:17.420 [2024-11-20 16:05:15.504158] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:17:17.420 [2024-11-20 16:05:15.504244] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:17.420 pt2 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:17.420 "name": "raid_bdev1", 00:17:17.420 "uuid": "ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded", 00:17:17.420 "strip_size_kb": 0, 00:17:17.420 "state": "online", 00:17:17.420 "raid_level": "raid1", 00:17:17.420 "superblock": true, 00:17:17.420 "num_base_bdevs": 2, 00:17:17.420 "num_base_bdevs_discovered": 2, 00:17:17.420 "num_base_bdevs_operational": 2, 00:17:17.420 "base_bdevs_list": [ 00:17:17.420 { 00:17:17.420 "name": "pt1", 00:17:17.420 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:17.420 "is_configured": true, 00:17:17.420 "data_offset": 256, 00:17:17.420 "data_size": 7936 00:17:17.420 }, 00:17:17.420 { 00:17:17.420 "name": "pt2", 00:17:17.420 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:17.420 "is_configured": true, 00:17:17.420 "data_offset": 256, 00:17:17.420 "data_size": 7936 00:17:17.420 } 00:17:17.420 ] 00:17:17.420 }' 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:17.420 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:17:17.682 [2024-11-20 16:05:15.831753] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:17:17.682 "name": "raid_bdev1", 00:17:17.682 "aliases": [ 00:17:17.682 "ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded" 00:17:17.682 ], 00:17:17.682 "product_name": "Raid Volume", 00:17:17.682 "block_size": 4096, 00:17:17.682 "num_blocks": 7936, 00:17:17.682 "uuid": "ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded", 00:17:17.682 "md_size": 32, 00:17:17.682 "md_interleave": false, 00:17:17.682 "dif_type": 0, 00:17:17.682 "assigned_rate_limits": { 00:17:17.682 "rw_ios_per_sec": 0, 00:17:17.682 "rw_mbytes_per_sec": 0, 00:17:17.682 "r_mbytes_per_sec": 0, 00:17:17.682 "w_mbytes_per_sec": 0 00:17:17.682 }, 00:17:17.682 "claimed": false, 00:17:17.682 "zoned": false, 00:17:17.682 "supported_io_types": { 00:17:17.682 "read": true, 00:17:17.682 "write": true, 00:17:17.682 "unmap": false, 00:17:17.682 "flush": false, 00:17:17.682 "reset": true, 00:17:17.682 "nvme_admin": false, 00:17:17.682 "nvme_io": false, 00:17:17.682 "nvme_io_md": false, 00:17:17.682 "write_zeroes": true, 00:17:17.682 "zcopy": false, 00:17:17.682 "get_zone_info": false, 00:17:17.682 "zone_management": false, 00:17:17.682 "zone_append": false, 00:17:17.682 "compare": false, 00:17:17.682 "compare_and_write": false, 00:17:17.682 "abort": false, 00:17:17.682 "seek_hole": false, 00:17:17.682 "seek_data": false, 00:17:17.682 "copy": false, 00:17:17.682 "nvme_iov_md": false 00:17:17.682 }, 00:17:17.682 "memory_domains": [ 00:17:17.682 { 00:17:17.682 "dma_device_id": "system", 00:17:17.682 "dma_device_type": 1 00:17:17.682 }, 00:17:17.682 { 00:17:17.682 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:17.682 "dma_device_type": 2 00:17:17.682 }, 00:17:17.682 { 00:17:17.682 "dma_device_id": "system", 00:17:17.682 "dma_device_type": 1 00:17:17.682 }, 00:17:17.682 { 00:17:17.682 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:17.682 "dma_device_type": 2 00:17:17.682 } 00:17:17.682 ], 00:17:17.682 "driver_specific": { 00:17:17.682 "raid": { 00:17:17.682 "uuid": "ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded", 00:17:17.682 "strip_size_kb": 0, 00:17:17.682 "state": "online", 00:17:17.682 "raid_level": "raid1", 00:17:17.682 "superblock": true, 00:17:17.682 "num_base_bdevs": 2, 00:17:17.682 "num_base_bdevs_discovered": 2, 00:17:17.682 "num_base_bdevs_operational": 2, 00:17:17.682 "base_bdevs_list": [ 00:17:17.682 { 00:17:17.682 "name": "pt1", 00:17:17.682 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:17.682 "is_configured": true, 00:17:17.682 "data_offset": 256, 00:17:17.682 "data_size": 7936 00:17:17.682 }, 00:17:17.682 { 00:17:17.682 "name": "pt2", 00:17:17.682 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:17.682 "is_configured": true, 00:17:17.682 "data_offset": 256, 00:17:17.682 "data_size": 7936 00:17:17.682 } 00:17:17.682 ] 00:17:17.682 } 00:17:17.682 } 00:17:17.682 }' 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:17:17.682 pt2' 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:17.682 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:17:17.944 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.944 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.944 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.944 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:17:17.944 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:17:17.944 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:17.944 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:17:17.944 16:05:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:17.944 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.944 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.944 16:05:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.944 [2024-11-20 16:05:16.011781] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # '[' ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded '!=' ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded ']' 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.944 [2024-11-20 16:05:16.043494] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.944 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:17.944 "name": "raid_bdev1", 00:17:17.945 "uuid": "ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded", 00:17:17.945 "strip_size_kb": 0, 00:17:17.945 "state": "online", 00:17:17.945 "raid_level": "raid1", 00:17:17.945 "superblock": true, 00:17:17.945 "num_base_bdevs": 2, 00:17:17.945 "num_base_bdevs_discovered": 1, 00:17:17.945 "num_base_bdevs_operational": 1, 00:17:17.945 "base_bdevs_list": [ 00:17:17.945 { 00:17:17.945 "name": null, 00:17:17.945 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:17.945 "is_configured": false, 00:17:17.945 "data_offset": 0, 00:17:17.945 "data_size": 7936 00:17:17.945 }, 00:17:17.945 { 00:17:17.945 "name": "pt2", 00:17:17.945 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:17.945 "is_configured": true, 00:17:17.945 "data_offset": 256, 00:17:17.945 "data_size": 7936 00:17:17.945 } 00:17:17.945 ] 00:17:17.945 }' 00:17:17.945 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:17.945 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.206 [2024-11-20 16:05:16.375570] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:18.206 [2024-11-20 16:05:16.375710] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:18.206 [2024-11-20 16:05:16.375850] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:18.206 [2024-11-20 16:05:16.375945] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:18.206 [2024-11-20 16:05:16.376377] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@519 -- # i=1 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.206 [2024-11-20 16:05:16.431575] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:18.206 [2024-11-20 16:05:16.431734] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:18.206 [2024-11-20 16:05:16.431775] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:17:18.206 [2024-11-20 16:05:16.431839] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:18.206 [2024-11-20 16:05:16.433841] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:18.206 [2024-11-20 16:05:16.433946] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:18.206 [2024-11-20 16:05:16.434036] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:17:18.206 [2024-11-20 16:05:16.434100] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:18.206 [2024-11-20 16:05:16.434208] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:17:18.206 [2024-11-20 16:05:16.434237] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:17:18.206 [2024-11-20 16:05:16.434363] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:17:18.206 [2024-11-20 16:05:16.434486] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:17:18.206 [2024-11-20 16:05:16.434547] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:17:18.206 [2024-11-20 16:05:16.434662] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:18.206 pt2 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:18.206 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:18.207 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:18.207 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:18.207 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:18.207 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:18.207 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:18.207 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:18.207 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:18.207 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:18.207 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.207 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.207 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:18.207 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.465 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:18.465 "name": "raid_bdev1", 00:17:18.465 "uuid": "ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded", 00:17:18.465 "strip_size_kb": 0, 00:17:18.465 "state": "online", 00:17:18.465 "raid_level": "raid1", 00:17:18.465 "superblock": true, 00:17:18.465 "num_base_bdevs": 2, 00:17:18.465 "num_base_bdevs_discovered": 1, 00:17:18.465 "num_base_bdevs_operational": 1, 00:17:18.465 "base_bdevs_list": [ 00:17:18.465 { 00:17:18.465 "name": null, 00:17:18.465 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:18.465 "is_configured": false, 00:17:18.465 "data_offset": 256, 00:17:18.465 "data_size": 7936 00:17:18.465 }, 00:17:18.465 { 00:17:18.465 "name": "pt2", 00:17:18.465 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:18.465 "is_configured": true, 00:17:18.465 "data_offset": 256, 00:17:18.465 "data_size": 7936 00:17:18.465 } 00:17:18.465 ] 00:17:18.465 }' 00:17:18.465 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:18.465 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.723 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:18.723 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.723 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.723 [2024-11-20 16:05:16.755625] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:18.723 [2024-11-20 16:05:16.755651] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:18.723 [2024-11-20 16:05:16.755720] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:18.723 [2024-11-20 16:05:16.755769] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:18.723 [2024-11-20 16:05:16.755778] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:17:18.723 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.723 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:18.723 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.723 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.723 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:17:18.723 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.723 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:17:18.723 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:17:18.723 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.724 [2024-11-20 16:05:16.799654] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:17:18.724 [2024-11-20 16:05:16.799717] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:18.724 [2024-11-20 16:05:16.799734] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:17:18.724 [2024-11-20 16:05:16.799743] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:18.724 [2024-11-20 16:05:16.801730] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:18.724 [2024-11-20 16:05:16.801768] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:17:18.724 [2024-11-20 16:05:16.801818] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:17:18.724 [2024-11-20 16:05:16.801857] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:17:18.724 [2024-11-20 16:05:16.801973] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:17:18.724 [2024-11-20 16:05:16.801983] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:18.724 [2024-11-20 16:05:16.802000] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:17:18.724 [2024-11-20 16:05:16.802053] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:18.724 [2024-11-20 16:05:16.802117] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:17:18.724 [2024-11-20 16:05:16.802126] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:17:18.724 [2024-11-20 16:05:16.802192] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:17:18.724 [2024-11-20 16:05:16.802291] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:17:18.724 [2024-11-20 16:05:16.802302] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:17:18.724 [2024-11-20 16:05:16.802394] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:18.724 pt1 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:18.724 "name": "raid_bdev1", 00:17:18.724 "uuid": "ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded", 00:17:18.724 "strip_size_kb": 0, 00:17:18.724 "state": "online", 00:17:18.724 "raid_level": "raid1", 00:17:18.724 "superblock": true, 00:17:18.724 "num_base_bdevs": 2, 00:17:18.724 "num_base_bdevs_discovered": 1, 00:17:18.724 "num_base_bdevs_operational": 1, 00:17:18.724 "base_bdevs_list": [ 00:17:18.724 { 00:17:18.724 "name": null, 00:17:18.724 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:18.724 "is_configured": false, 00:17:18.724 "data_offset": 256, 00:17:18.724 "data_size": 7936 00:17:18.724 }, 00:17:18.724 { 00:17:18.724 "name": "pt2", 00:17:18.724 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:18.724 "is_configured": true, 00:17:18.724 "data_offset": 256, 00:17:18.724 "data_size": 7936 00:17:18.724 } 00:17:18.724 ] 00:17:18.724 }' 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:18.724 16:05:16 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.982 [2024-11-20 16:05:17.152003] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # '[' ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded '!=' ab46136f-cb0d-4e2b-8ecb-6cdc341b8ded ']' 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@563 -- # killprocess 84966 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@954 -- # '[' -z 84966 ']' 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@958 -- # kill -0 84966 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@959 -- # uname 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 84966 00:17:18.982 killing process with pid 84966 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 84966' 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@973 -- # kill 84966 00:17:18.982 [2024-11-20 16:05:17.205204] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:18.982 16:05:17 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@978 -- # wait 84966 00:17:18.982 [2024-11-20 16:05:17.205285] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:18.982 [2024-11-20 16:05:17.205334] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:18.982 [2024-11-20 16:05:17.205352] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:17:19.240 [2024-11-20 16:05:17.349620] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:20.180 16:05:18 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@565 -- # return 0 00:17:20.180 00:17:20.180 real 0m4.653s 00:17:20.180 user 0m7.041s 00:17:20.180 sys 0m0.746s 00:17:20.180 16:05:18 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:20.180 ************************************ 00:17:20.180 END TEST raid_superblock_test_md_separate 00:17:20.180 ************************************ 00:17:20.180 16:05:18 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.180 16:05:18 bdev_raid -- bdev/bdev_raid.sh@1006 -- # '[' true = true ']' 00:17:20.180 16:05:18 bdev_raid -- bdev/bdev_raid.sh@1007 -- # run_test raid_rebuild_test_sb_md_separate raid_rebuild_test raid1 2 true false true 00:17:20.180 16:05:18 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:17:20.180 16:05:18 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:20.180 16:05:18 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:20.180 ************************************ 00:17:20.180 START TEST raid_rebuild_test_sb_md_separate 00:17:20.180 ************************************ 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # local verify=true 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # local strip_size 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@577 -- # local create_arg 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@579 -- # local data_offset 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:17:20.180 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@597 -- # raid_pid=85278 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@598 -- # waitforlisten 85278 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@835 -- # '[' -z 85278 ']' 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.180 16:05:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:17:20.180 I/O size of 3145728 is greater than zero copy threshold (65536). 00:17:20.180 Zero copy mechanism will not be used. 00:17:20.180 [2024-11-20 16:05:18.192424] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:17:20.180 [2024-11-20 16:05:18.192524] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85278 ] 00:17:20.180 [2024-11-20 16:05:18.347270] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:20.452 [2024-11-20 16:05:18.452293] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:20.452 [2024-11-20 16:05:18.591038] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:20.452 [2024-11-20 16:05:18.591076] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # return 0 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1_malloc 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.021 BaseBdev1_malloc 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.021 [2024-11-20 16:05:19.079655] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:17:21.021 [2024-11-20 16:05:19.079724] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:21.021 [2024-11-20 16:05:19.079744] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:17:21.021 [2024-11-20 16:05:19.079756] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:21.021 [2024-11-20 16:05:19.081685] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:21.021 [2024-11-20 16:05:19.081719] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:17:21.021 BaseBdev1 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2_malloc 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.021 BaseBdev2_malloc 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.021 [2024-11-20 16:05:19.120884] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:17:21.021 [2024-11-20 16:05:19.121058] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:21.021 [2024-11-20 16:05:19.121082] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:17:21.021 [2024-11-20 16:05:19.121094] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:21.021 [2024-11-20 16:05:19.123056] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:21.021 [2024-11-20 16:05:19.123092] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:17:21.021 BaseBdev2 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b spare_malloc 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.021 spare_malloc 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.021 spare_delay 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.021 [2024-11-20 16:05:19.177344] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:21.021 [2024-11-20 16:05:19.177397] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:21.021 [2024-11-20 16:05:19.177415] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:17:21.021 [2024-11-20 16:05:19.177426] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:21.021 [2024-11-20 16:05:19.179320] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:21.021 [2024-11-20 16:05:19.179356] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:21.021 spare 00:17:21.021 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.022 [2024-11-20 16:05:19.185386] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:21.022 [2024-11-20 16:05:19.187192] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:21.022 [2024-11-20 16:05:19.187350] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:17:21.022 [2024-11-20 16:05:19.187363] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:17:21.022 [2024-11-20 16:05:19.187435] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:17:21.022 [2024-11-20 16:05:19.187547] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:17:21.022 [2024-11-20 16:05:19.187557] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:17:21.022 [2024-11-20 16:05:19.187646] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:21.022 "name": "raid_bdev1", 00:17:21.022 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:21.022 "strip_size_kb": 0, 00:17:21.022 "state": "online", 00:17:21.022 "raid_level": "raid1", 00:17:21.022 "superblock": true, 00:17:21.022 "num_base_bdevs": 2, 00:17:21.022 "num_base_bdevs_discovered": 2, 00:17:21.022 "num_base_bdevs_operational": 2, 00:17:21.022 "base_bdevs_list": [ 00:17:21.022 { 00:17:21.022 "name": "BaseBdev1", 00:17:21.022 "uuid": "2011c49a-f3a6-5993-bd9c-ca47e1ce55a1", 00:17:21.022 "is_configured": true, 00:17:21.022 "data_offset": 256, 00:17:21.022 "data_size": 7936 00:17:21.022 }, 00:17:21.022 { 00:17:21.022 "name": "BaseBdev2", 00:17:21.022 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:21.022 "is_configured": true, 00:17:21.022 "data_offset": 256, 00:17:21.022 "data_size": 7936 00:17:21.022 } 00:17:21.022 ] 00:17:21.022 }' 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:21.022 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.279 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:21.279 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:17:21.279 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:21.279 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.279 [2024-11-20 16:05:19.525888] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:17:21.540 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:17:21.540 [2024-11-20 16:05:19.773594] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:17:21.801 /dev/nbd0 00:17:21.801 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:17:21.801 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:17:21.801 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:17:21.801 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:17:21.801 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:17:21.801 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:17:21.801 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:17:21.801 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:17:21.801 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:17:21.801 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:17:21.801 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:21.801 1+0 records in 00:17:21.801 1+0 records out 00:17:21.801 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000594852 s, 6.9 MB/s 00:17:21.801 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:21.801 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:17:21.801 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:21.801 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:17:21.801 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:17:21.802 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:21.802 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:17:21.802 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:17:21.802 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:17:21.802 16:05:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:17:22.373 7936+0 records in 00:17:22.373 7936+0 records out 00:17:22.373 32505856 bytes (33 MB, 31 MiB) copied, 0.740222 s, 43.9 MB/s 00:17:22.373 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:17:22.373 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:17:22.373 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:17:22.373 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:22.373 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:17:22.373 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:22.373 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:17:22.632 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:22.632 [2024-11-20 16:05:20.778382] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:22.632 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:22.633 [2024-11-20 16:05:20.787754] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:22.633 "name": "raid_bdev1", 00:17:22.633 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:22.633 "strip_size_kb": 0, 00:17:22.633 "state": "online", 00:17:22.633 "raid_level": "raid1", 00:17:22.633 "superblock": true, 00:17:22.633 "num_base_bdevs": 2, 00:17:22.633 "num_base_bdevs_discovered": 1, 00:17:22.633 "num_base_bdevs_operational": 1, 00:17:22.633 "base_bdevs_list": [ 00:17:22.633 { 00:17:22.633 "name": null, 00:17:22.633 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:22.633 "is_configured": false, 00:17:22.633 "data_offset": 0, 00:17:22.633 "data_size": 7936 00:17:22.633 }, 00:17:22.633 { 00:17:22.633 "name": "BaseBdev2", 00:17:22.633 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:22.633 "is_configured": true, 00:17:22.633 "data_offset": 256, 00:17:22.633 "data_size": 7936 00:17:22.633 } 00:17:22.633 ] 00:17:22.633 }' 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:22.633 16:05:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:22.891 16:05:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:22.891 16:05:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:22.891 16:05:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:22.891 [2024-11-20 16:05:21.111882] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:22.891 [2024-11-20 16:05:21.121784] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d260 00:17:22.891 16:05:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:22.891 16:05:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@647 -- # sleep 1 00:17:22.891 [2024-11-20 16:05:21.123688] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:24.270 "name": "raid_bdev1", 00:17:24.270 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:24.270 "strip_size_kb": 0, 00:17:24.270 "state": "online", 00:17:24.270 "raid_level": "raid1", 00:17:24.270 "superblock": true, 00:17:24.270 "num_base_bdevs": 2, 00:17:24.270 "num_base_bdevs_discovered": 2, 00:17:24.270 "num_base_bdevs_operational": 2, 00:17:24.270 "process": { 00:17:24.270 "type": "rebuild", 00:17:24.270 "target": "spare", 00:17:24.270 "progress": { 00:17:24.270 "blocks": 2560, 00:17:24.270 "percent": 32 00:17:24.270 } 00:17:24.270 }, 00:17:24.270 "base_bdevs_list": [ 00:17:24.270 { 00:17:24.270 "name": "spare", 00:17:24.270 "uuid": "8ecdeb83-6fef-5b9c-ac3d-1c7d40d43405", 00:17:24.270 "is_configured": true, 00:17:24.270 "data_offset": 256, 00:17:24.270 "data_size": 7936 00:17:24.270 }, 00:17:24.270 { 00:17:24.270 "name": "BaseBdev2", 00:17:24.270 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:24.270 "is_configured": true, 00:17:24.270 "data_offset": 256, 00:17:24.270 "data_size": 7936 00:17:24.270 } 00:17:24.270 ] 00:17:24.270 }' 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:24.270 [2024-11-20 16:05:22.229953] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:24.270 [2024-11-20 16:05:22.330146] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:24.270 [2024-11-20 16:05:22.330417] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:24.270 [2024-11-20 16:05:22.330643] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:24.270 [2024-11-20 16:05:22.330698] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:24.270 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:24.271 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:24.271 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:24.271 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:24.271 "name": "raid_bdev1", 00:17:24.271 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:24.271 "strip_size_kb": 0, 00:17:24.271 "state": "online", 00:17:24.271 "raid_level": "raid1", 00:17:24.271 "superblock": true, 00:17:24.271 "num_base_bdevs": 2, 00:17:24.271 "num_base_bdevs_discovered": 1, 00:17:24.271 "num_base_bdevs_operational": 1, 00:17:24.271 "base_bdevs_list": [ 00:17:24.271 { 00:17:24.271 "name": null, 00:17:24.271 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:24.271 "is_configured": false, 00:17:24.271 "data_offset": 0, 00:17:24.271 "data_size": 7936 00:17:24.271 }, 00:17:24.271 { 00:17:24.271 "name": "BaseBdev2", 00:17:24.271 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:24.271 "is_configured": true, 00:17:24.271 "data_offset": 256, 00:17:24.271 "data_size": 7936 00:17:24.271 } 00:17:24.271 ] 00:17:24.271 }' 00:17:24.271 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:24.271 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:24.531 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:24.531 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:24.531 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:24.531 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:24.531 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:24.531 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:24.531 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:24.531 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:24.531 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:24.531 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:24.531 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:24.531 "name": "raid_bdev1", 00:17:24.531 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:24.531 "strip_size_kb": 0, 00:17:24.531 "state": "online", 00:17:24.531 "raid_level": "raid1", 00:17:24.531 "superblock": true, 00:17:24.531 "num_base_bdevs": 2, 00:17:24.531 "num_base_bdevs_discovered": 1, 00:17:24.531 "num_base_bdevs_operational": 1, 00:17:24.531 "base_bdevs_list": [ 00:17:24.531 { 00:17:24.531 "name": null, 00:17:24.531 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:24.531 "is_configured": false, 00:17:24.531 "data_offset": 0, 00:17:24.531 "data_size": 7936 00:17:24.531 }, 00:17:24.532 { 00:17:24.532 "name": "BaseBdev2", 00:17:24.532 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:24.532 "is_configured": true, 00:17:24.532 "data_offset": 256, 00:17:24.532 "data_size": 7936 00:17:24.532 } 00:17:24.532 ] 00:17:24.532 }' 00:17:24.532 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:24.532 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:24.532 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:24.532 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:24.532 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:24.532 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:24.532 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:24.532 [2024-11-20 16:05:22.745458] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:24.532 [2024-11-20 16:05:22.754913] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d330 00:17:24.532 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:24.532 16:05:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@663 -- # sleep 1 00:17:24.532 [2024-11-20 16:05:22.756918] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:25.921 "name": "raid_bdev1", 00:17:25.921 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:25.921 "strip_size_kb": 0, 00:17:25.921 "state": "online", 00:17:25.921 "raid_level": "raid1", 00:17:25.921 "superblock": true, 00:17:25.921 "num_base_bdevs": 2, 00:17:25.921 "num_base_bdevs_discovered": 2, 00:17:25.921 "num_base_bdevs_operational": 2, 00:17:25.921 "process": { 00:17:25.921 "type": "rebuild", 00:17:25.921 "target": "spare", 00:17:25.921 "progress": { 00:17:25.921 "blocks": 2560, 00:17:25.921 "percent": 32 00:17:25.921 } 00:17:25.921 }, 00:17:25.921 "base_bdevs_list": [ 00:17:25.921 { 00:17:25.921 "name": "spare", 00:17:25.921 "uuid": "8ecdeb83-6fef-5b9c-ac3d-1c7d40d43405", 00:17:25.921 "is_configured": true, 00:17:25.921 "data_offset": 256, 00:17:25.921 "data_size": 7936 00:17:25.921 }, 00:17:25.921 { 00:17:25.921 "name": "BaseBdev2", 00:17:25.921 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:25.921 "is_configured": true, 00:17:25.921 "data_offset": 256, 00:17:25.921 "data_size": 7936 00:17:25.921 } 00:17:25.921 ] 00:17:25.921 }' 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:17:25.921 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@706 -- # local timeout=577 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:25.921 "name": "raid_bdev1", 00:17:25.921 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:25.921 "strip_size_kb": 0, 00:17:25.921 "state": "online", 00:17:25.921 "raid_level": "raid1", 00:17:25.921 "superblock": true, 00:17:25.921 "num_base_bdevs": 2, 00:17:25.921 "num_base_bdevs_discovered": 2, 00:17:25.921 "num_base_bdevs_operational": 2, 00:17:25.921 "process": { 00:17:25.921 "type": "rebuild", 00:17:25.921 "target": "spare", 00:17:25.921 "progress": { 00:17:25.921 "blocks": 2816, 00:17:25.921 "percent": 35 00:17:25.921 } 00:17:25.921 }, 00:17:25.921 "base_bdevs_list": [ 00:17:25.921 { 00:17:25.921 "name": "spare", 00:17:25.921 "uuid": "8ecdeb83-6fef-5b9c-ac3d-1c7d40d43405", 00:17:25.921 "is_configured": true, 00:17:25.921 "data_offset": 256, 00:17:25.921 "data_size": 7936 00:17:25.921 }, 00:17:25.921 { 00:17:25.921 "name": "BaseBdev2", 00:17:25.921 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:25.921 "is_configured": true, 00:17:25.921 "data_offset": 256, 00:17:25.921 "data_size": 7936 00:17:25.921 } 00:17:25.921 ] 00:17:25.921 }' 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:25.921 16:05:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:17:26.932 16:05:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:26.932 16:05:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:26.932 16:05:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:26.932 16:05:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:26.932 16:05:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:26.932 16:05:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:26.932 16:05:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:26.932 16:05:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:26.932 16:05:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:26.932 16:05:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:26.932 16:05:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:26.932 16:05:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:26.932 "name": "raid_bdev1", 00:17:26.932 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:26.932 "strip_size_kb": 0, 00:17:26.932 "state": "online", 00:17:26.932 "raid_level": "raid1", 00:17:26.932 "superblock": true, 00:17:26.932 "num_base_bdevs": 2, 00:17:26.932 "num_base_bdevs_discovered": 2, 00:17:26.932 "num_base_bdevs_operational": 2, 00:17:26.932 "process": { 00:17:26.932 "type": "rebuild", 00:17:26.932 "target": "spare", 00:17:26.932 "progress": { 00:17:26.932 "blocks": 5376, 00:17:26.932 "percent": 67 00:17:26.932 } 00:17:26.932 }, 00:17:26.932 "base_bdevs_list": [ 00:17:26.932 { 00:17:26.932 "name": "spare", 00:17:26.932 "uuid": "8ecdeb83-6fef-5b9c-ac3d-1c7d40d43405", 00:17:26.932 "is_configured": true, 00:17:26.932 "data_offset": 256, 00:17:26.932 "data_size": 7936 00:17:26.932 }, 00:17:26.932 { 00:17:26.932 "name": "BaseBdev2", 00:17:26.932 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:26.933 "is_configured": true, 00:17:26.933 "data_offset": 256, 00:17:26.933 "data_size": 7936 00:17:26.933 } 00:17:26.933 ] 00:17:26.933 }' 00:17:26.933 16:05:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:26.933 16:05:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:26.933 16:05:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:26.933 16:05:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:26.933 16:05:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:17:27.876 [2024-11-20 16:05:25.871947] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:17:27.876 [2024-11-20 16:05:25.872028] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:17:27.876 [2024-11-20 16:05:25.872136] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:27.876 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:27.876 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:27.876 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:27.876 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:27.876 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:27.876 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:27.876 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:27.876 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:27.876 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:27.876 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:27.876 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:27.876 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:27.876 "name": "raid_bdev1", 00:17:27.876 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:27.876 "strip_size_kb": 0, 00:17:27.876 "state": "online", 00:17:27.876 "raid_level": "raid1", 00:17:27.876 "superblock": true, 00:17:27.876 "num_base_bdevs": 2, 00:17:27.876 "num_base_bdevs_discovered": 2, 00:17:27.876 "num_base_bdevs_operational": 2, 00:17:27.876 "base_bdevs_list": [ 00:17:27.876 { 00:17:27.876 "name": "spare", 00:17:27.876 "uuid": "8ecdeb83-6fef-5b9c-ac3d-1c7d40d43405", 00:17:27.876 "is_configured": true, 00:17:27.876 "data_offset": 256, 00:17:27.876 "data_size": 7936 00:17:27.876 }, 00:17:27.876 { 00:17:27.876 "name": "BaseBdev2", 00:17:27.876 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:27.876 "is_configured": true, 00:17:27.876 "data_offset": 256, 00:17:27.876 "data_size": 7936 00:17:27.876 } 00:17:27.876 ] 00:17:27.876 }' 00:17:27.876 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@709 -- # break 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:28.138 "name": "raid_bdev1", 00:17:28.138 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:28.138 "strip_size_kb": 0, 00:17:28.138 "state": "online", 00:17:28.138 "raid_level": "raid1", 00:17:28.138 "superblock": true, 00:17:28.138 "num_base_bdevs": 2, 00:17:28.138 "num_base_bdevs_discovered": 2, 00:17:28.138 "num_base_bdevs_operational": 2, 00:17:28.138 "base_bdevs_list": [ 00:17:28.138 { 00:17:28.138 "name": "spare", 00:17:28.138 "uuid": "8ecdeb83-6fef-5b9c-ac3d-1c7d40d43405", 00:17:28.138 "is_configured": true, 00:17:28.138 "data_offset": 256, 00:17:28.138 "data_size": 7936 00:17:28.138 }, 00:17:28.138 { 00:17:28.138 "name": "BaseBdev2", 00:17:28.138 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:28.138 "is_configured": true, 00:17:28.138 "data_offset": 256, 00:17:28.138 "data_size": 7936 00:17:28.138 } 00:17:28.138 ] 00:17:28.138 }' 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:28.138 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:28.138 "name": "raid_bdev1", 00:17:28.138 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:28.138 "strip_size_kb": 0, 00:17:28.138 "state": "online", 00:17:28.138 "raid_level": "raid1", 00:17:28.138 "superblock": true, 00:17:28.138 "num_base_bdevs": 2, 00:17:28.138 "num_base_bdevs_discovered": 2, 00:17:28.139 "num_base_bdevs_operational": 2, 00:17:28.139 "base_bdevs_list": [ 00:17:28.139 { 00:17:28.139 "name": "spare", 00:17:28.139 "uuid": "8ecdeb83-6fef-5b9c-ac3d-1c7d40d43405", 00:17:28.139 "is_configured": true, 00:17:28.139 "data_offset": 256, 00:17:28.139 "data_size": 7936 00:17:28.139 }, 00:17:28.139 { 00:17:28.139 "name": "BaseBdev2", 00:17:28.139 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:28.139 "is_configured": true, 00:17:28.139 "data_offset": 256, 00:17:28.139 "data_size": 7936 00:17:28.139 } 00:17:28.139 ] 00:17:28.139 }' 00:17:28.139 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:28.139 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:28.400 [2024-11-20 16:05:26.574286] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:28.400 [2024-11-20 16:05:26.574312] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:28.400 [2024-11-20 16:05:26.574380] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:28.400 [2024-11-20 16:05:26.574450] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:28.400 [2024-11-20 16:05:26.574461] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # jq length 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:17:28.400 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:28.401 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:17:28.662 /dev/nbd0 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:28.662 1+0 records in 00:17:28.662 1+0 records out 00:17:28.662 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000235421 s, 17.4 MB/s 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:28.662 16:05:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:17:28.924 /dev/nbd1 00:17:28.924 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:17:28.924 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:17:28.924 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:17:28.924 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:17:28.924 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:17:28.924 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:17:28.924 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:17:28.924 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:17:28.924 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:17:28.924 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:17:28.924 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:28.924 1+0 records in 00:17:28.924 1+0 records out 00:17:28.924 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000400702 s, 10.2 MB/s 00:17:28.924 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:28.924 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:17:28.924 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:28.924 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:17:28.925 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:17:28.925 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:28.925 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:28.925 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:17:29.186 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:17:29.186 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:17:29.186 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:29.186 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:29.186 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:17:29.186 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:29.186 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:17:29.186 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:29.448 [2024-11-20 16:05:27.665011] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:29.448 [2024-11-20 16:05:27.665060] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:29.448 [2024-11-20 16:05:27.665081] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:17:29.448 [2024-11-20 16:05:27.665090] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:29.448 [2024-11-20 16:05:27.667105] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:29.448 [2024-11-20 16:05:27.667138] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:29.448 [2024-11-20 16:05:27.667198] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:17:29.448 [2024-11-20 16:05:27.667245] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:29.448 [2024-11-20 16:05:27.667366] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:29.448 spare 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:29.448 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:29.710 [2024-11-20 16:05:27.767441] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:17:29.710 [2024-11-20 16:05:27.767469] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:17:29.710 [2024-11-20 16:05:27.767573] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1b50 00:17:29.710 [2024-11-20 16:05:27.767750] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:17:29.710 [2024-11-20 16:05:27.767760] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:17:29.710 [2024-11-20 16:05:27.767882] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:29.710 "name": "raid_bdev1", 00:17:29.710 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:29.710 "strip_size_kb": 0, 00:17:29.710 "state": "online", 00:17:29.710 "raid_level": "raid1", 00:17:29.710 "superblock": true, 00:17:29.710 "num_base_bdevs": 2, 00:17:29.710 "num_base_bdevs_discovered": 2, 00:17:29.710 "num_base_bdevs_operational": 2, 00:17:29.710 "base_bdevs_list": [ 00:17:29.710 { 00:17:29.710 "name": "spare", 00:17:29.710 "uuid": "8ecdeb83-6fef-5b9c-ac3d-1c7d40d43405", 00:17:29.710 "is_configured": true, 00:17:29.710 "data_offset": 256, 00:17:29.710 "data_size": 7936 00:17:29.710 }, 00:17:29.710 { 00:17:29.710 "name": "BaseBdev2", 00:17:29.710 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:29.710 "is_configured": true, 00:17:29.710 "data_offset": 256, 00:17:29.710 "data_size": 7936 00:17:29.710 } 00:17:29.710 ] 00:17:29.710 }' 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:29.710 16:05:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:29.972 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:29.972 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:29.972 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:29.972 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:29.972 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:29.972 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:29.972 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:29.972 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:29.973 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:29.973 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:29.973 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:29.973 "name": "raid_bdev1", 00:17:29.973 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:29.973 "strip_size_kb": 0, 00:17:29.973 "state": "online", 00:17:29.973 "raid_level": "raid1", 00:17:29.973 "superblock": true, 00:17:29.973 "num_base_bdevs": 2, 00:17:29.973 "num_base_bdevs_discovered": 2, 00:17:29.973 "num_base_bdevs_operational": 2, 00:17:29.973 "base_bdevs_list": [ 00:17:29.973 { 00:17:29.973 "name": "spare", 00:17:29.973 "uuid": "8ecdeb83-6fef-5b9c-ac3d-1c7d40d43405", 00:17:29.973 "is_configured": true, 00:17:29.973 "data_offset": 256, 00:17:29.973 "data_size": 7936 00:17:29.973 }, 00:17:29.973 { 00:17:29.973 "name": "BaseBdev2", 00:17:29.973 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:29.973 "is_configured": true, 00:17:29.973 "data_offset": 256, 00:17:29.973 "data_size": 7936 00:17:29.973 } 00:17:29.973 ] 00:17:29.973 }' 00:17:29.973 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:29.973 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:29.973 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:29.973 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:29.973 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:17:29.973 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:29.973 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:29.973 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:29.973 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:30.234 [2024-11-20 16:05:28.229167] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:30.234 "name": "raid_bdev1", 00:17:30.234 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:30.234 "strip_size_kb": 0, 00:17:30.234 "state": "online", 00:17:30.234 "raid_level": "raid1", 00:17:30.234 "superblock": true, 00:17:30.234 "num_base_bdevs": 2, 00:17:30.234 "num_base_bdevs_discovered": 1, 00:17:30.234 "num_base_bdevs_operational": 1, 00:17:30.234 "base_bdevs_list": [ 00:17:30.234 { 00:17:30.234 "name": null, 00:17:30.234 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:30.234 "is_configured": false, 00:17:30.234 "data_offset": 0, 00:17:30.234 "data_size": 7936 00:17:30.234 }, 00:17:30.234 { 00:17:30.234 "name": "BaseBdev2", 00:17:30.234 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:30.234 "is_configured": true, 00:17:30.234 "data_offset": 256, 00:17:30.234 "data_size": 7936 00:17:30.234 } 00:17:30.234 ] 00:17:30.234 }' 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:30.234 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:30.496 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:30.496 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:30.496 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:30.496 [2024-11-20 16:05:28.545259] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:30.496 [2024-11-20 16:05:28.545420] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:17:30.496 [2024-11-20 16:05:28.545436] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:17:30.496 [2024-11-20 16:05:28.545465] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:30.496 [2024-11-20 16:05:28.554426] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1c20 00:17:30.496 [2024-11-20 16:05:28.556383] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:30.496 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:30.496 16:05:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@757 -- # sleep 1 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:31.468 "name": "raid_bdev1", 00:17:31.468 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:31.468 "strip_size_kb": 0, 00:17:31.468 "state": "online", 00:17:31.468 "raid_level": "raid1", 00:17:31.468 "superblock": true, 00:17:31.468 "num_base_bdevs": 2, 00:17:31.468 "num_base_bdevs_discovered": 2, 00:17:31.468 "num_base_bdevs_operational": 2, 00:17:31.468 "process": { 00:17:31.468 "type": "rebuild", 00:17:31.468 "target": "spare", 00:17:31.468 "progress": { 00:17:31.468 "blocks": 2560, 00:17:31.468 "percent": 32 00:17:31.468 } 00:17:31.468 }, 00:17:31.468 "base_bdevs_list": [ 00:17:31.468 { 00:17:31.468 "name": "spare", 00:17:31.468 "uuid": "8ecdeb83-6fef-5b9c-ac3d-1c7d40d43405", 00:17:31.468 "is_configured": true, 00:17:31.468 "data_offset": 256, 00:17:31.468 "data_size": 7936 00:17:31.468 }, 00:17:31.468 { 00:17:31.468 "name": "BaseBdev2", 00:17:31.468 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:31.468 "is_configured": true, 00:17:31.468 "data_offset": 256, 00:17:31.468 "data_size": 7936 00:17:31.468 } 00:17:31.468 ] 00:17:31.468 }' 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:31.468 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:31.468 [2024-11-20 16:05:29.666390] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:31.730 [2024-11-20 16:05:29.762138] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:31.730 [2024-11-20 16:05:29.762197] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:31.730 [2024-11-20 16:05:29.762213] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:31.730 [2024-11-20 16:05:29.762230] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:31.730 "name": "raid_bdev1", 00:17:31.730 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:31.730 "strip_size_kb": 0, 00:17:31.730 "state": "online", 00:17:31.730 "raid_level": "raid1", 00:17:31.730 "superblock": true, 00:17:31.730 "num_base_bdevs": 2, 00:17:31.730 "num_base_bdevs_discovered": 1, 00:17:31.730 "num_base_bdevs_operational": 1, 00:17:31.730 "base_bdevs_list": [ 00:17:31.730 { 00:17:31.730 "name": null, 00:17:31.730 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:31.730 "is_configured": false, 00:17:31.730 "data_offset": 0, 00:17:31.730 "data_size": 7936 00:17:31.730 }, 00:17:31.730 { 00:17:31.730 "name": "BaseBdev2", 00:17:31.730 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:31.730 "is_configured": true, 00:17:31.730 "data_offset": 256, 00:17:31.730 "data_size": 7936 00:17:31.730 } 00:17:31.730 ] 00:17:31.730 }' 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:31.730 16:05:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:31.992 16:05:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:31.992 16:05:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:31.992 16:05:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:31.992 [2024-11-20 16:05:30.092225] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:31.992 [2024-11-20 16:05:30.092279] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:31.992 [2024-11-20 16:05:30.092302] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:17:31.992 [2024-11-20 16:05:30.092312] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:31.992 [2024-11-20 16:05:30.092541] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:31.992 [2024-11-20 16:05:30.092555] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:31.992 [2024-11-20 16:05:30.092607] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:17:31.992 [2024-11-20 16:05:30.092619] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:17:31.992 [2024-11-20 16:05:30.092629] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:17:31.992 [2024-11-20 16:05:30.092648] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:31.992 [2024-11-20 16:05:30.101411] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1cf0 00:17:31.992 spare 00:17:31.992 16:05:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:31.992 16:05:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@764 -- # sleep 1 00:17:31.992 [2024-11-20 16:05:30.103293] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:32.935 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:32.935 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:32.936 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:32.936 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:32.936 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:32.936 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:32.936 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:32.936 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:32.936 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:32.936 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:32.936 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:32.936 "name": "raid_bdev1", 00:17:32.936 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:32.936 "strip_size_kb": 0, 00:17:32.936 "state": "online", 00:17:32.936 "raid_level": "raid1", 00:17:32.936 "superblock": true, 00:17:32.936 "num_base_bdevs": 2, 00:17:32.936 "num_base_bdevs_discovered": 2, 00:17:32.936 "num_base_bdevs_operational": 2, 00:17:32.936 "process": { 00:17:32.936 "type": "rebuild", 00:17:32.936 "target": "spare", 00:17:32.936 "progress": { 00:17:32.936 "blocks": 2560, 00:17:32.936 "percent": 32 00:17:32.936 } 00:17:32.936 }, 00:17:32.936 "base_bdevs_list": [ 00:17:32.936 { 00:17:32.936 "name": "spare", 00:17:32.936 "uuid": "8ecdeb83-6fef-5b9c-ac3d-1c7d40d43405", 00:17:32.936 "is_configured": true, 00:17:32.936 "data_offset": 256, 00:17:32.936 "data_size": 7936 00:17:32.936 }, 00:17:32.936 { 00:17:32.936 "name": "BaseBdev2", 00:17:32.936 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:32.936 "is_configured": true, 00:17:32.936 "data_offset": 256, 00:17:32.936 "data_size": 7936 00:17:32.936 } 00:17:32.936 ] 00:17:32.936 }' 00:17:32.936 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:32.936 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:32.936 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:33.196 [2024-11-20 16:05:31.201781] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:33.196 [2024-11-20 16:05:31.208566] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:33.196 [2024-11-20 16:05:31.208736] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:33.196 [2024-11-20 16:05:31.208758] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:33.196 [2024-11-20 16:05:31.208766] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:33.196 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:33.196 "name": "raid_bdev1", 00:17:33.196 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:33.196 "strip_size_kb": 0, 00:17:33.196 "state": "online", 00:17:33.196 "raid_level": "raid1", 00:17:33.196 "superblock": true, 00:17:33.196 "num_base_bdevs": 2, 00:17:33.196 "num_base_bdevs_discovered": 1, 00:17:33.196 "num_base_bdevs_operational": 1, 00:17:33.196 "base_bdevs_list": [ 00:17:33.196 { 00:17:33.196 "name": null, 00:17:33.196 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:33.196 "is_configured": false, 00:17:33.196 "data_offset": 0, 00:17:33.196 "data_size": 7936 00:17:33.197 }, 00:17:33.197 { 00:17:33.197 "name": "BaseBdev2", 00:17:33.197 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:33.197 "is_configured": true, 00:17:33.197 "data_offset": 256, 00:17:33.197 "data_size": 7936 00:17:33.197 } 00:17:33.197 ] 00:17:33.197 }' 00:17:33.197 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:33.197 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:33.457 "name": "raid_bdev1", 00:17:33.457 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:33.457 "strip_size_kb": 0, 00:17:33.457 "state": "online", 00:17:33.457 "raid_level": "raid1", 00:17:33.457 "superblock": true, 00:17:33.457 "num_base_bdevs": 2, 00:17:33.457 "num_base_bdevs_discovered": 1, 00:17:33.457 "num_base_bdevs_operational": 1, 00:17:33.457 "base_bdevs_list": [ 00:17:33.457 { 00:17:33.457 "name": null, 00:17:33.457 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:33.457 "is_configured": false, 00:17:33.457 "data_offset": 0, 00:17:33.457 "data_size": 7936 00:17:33.457 }, 00:17:33.457 { 00:17:33.457 "name": "BaseBdev2", 00:17:33.457 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:33.457 "is_configured": true, 00:17:33.457 "data_offset": 256, 00:17:33.457 "data_size": 7936 00:17:33.457 } 00:17:33.457 ] 00:17:33.457 }' 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:33.457 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:33.458 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:17:33.458 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:33.458 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:33.458 [2024-11-20 16:05:31.650758] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:17:33.458 [2024-11-20 16:05:31.650802] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:33.458 [2024-11-20 16:05:31.650834] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:17:33.458 [2024-11-20 16:05:31.650843] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:33.458 [2024-11-20 16:05:31.651032] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:33.458 [2024-11-20 16:05:31.651044] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:17:33.458 [2024-11-20 16:05:31.651089] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:17:33.458 [2024-11-20 16:05:31.651100] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:17:33.458 [2024-11-20 16:05:31.651109] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:17:33.458 [2024-11-20 16:05:31.651118] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:17:33.458 BaseBdev1 00:17:33.458 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:33.458 16:05:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@775 -- # sleep 1 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:34.835 "name": "raid_bdev1", 00:17:34.835 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:34.835 "strip_size_kb": 0, 00:17:34.835 "state": "online", 00:17:34.835 "raid_level": "raid1", 00:17:34.835 "superblock": true, 00:17:34.835 "num_base_bdevs": 2, 00:17:34.835 "num_base_bdevs_discovered": 1, 00:17:34.835 "num_base_bdevs_operational": 1, 00:17:34.835 "base_bdevs_list": [ 00:17:34.835 { 00:17:34.835 "name": null, 00:17:34.835 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:34.835 "is_configured": false, 00:17:34.835 "data_offset": 0, 00:17:34.835 "data_size": 7936 00:17:34.835 }, 00:17:34.835 { 00:17:34.835 "name": "BaseBdev2", 00:17:34.835 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:34.835 "is_configured": true, 00:17:34.835 "data_offset": 256, 00:17:34.835 "data_size": 7936 00:17:34.835 } 00:17:34.835 ] 00:17:34.835 }' 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:34.835 16:05:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:34.835 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:34.835 "name": "raid_bdev1", 00:17:34.835 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:34.835 "strip_size_kb": 0, 00:17:34.835 "state": "online", 00:17:34.835 "raid_level": "raid1", 00:17:34.835 "superblock": true, 00:17:34.835 "num_base_bdevs": 2, 00:17:34.835 "num_base_bdevs_discovered": 1, 00:17:34.835 "num_base_bdevs_operational": 1, 00:17:34.835 "base_bdevs_list": [ 00:17:34.835 { 00:17:34.835 "name": null, 00:17:34.835 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:34.835 "is_configured": false, 00:17:34.835 "data_offset": 0, 00:17:34.835 "data_size": 7936 00:17:34.835 }, 00:17:34.835 { 00:17:34.835 "name": "BaseBdev2", 00:17:34.835 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:34.835 "is_configured": true, 00:17:34.835 "data_offset": 256, 00:17:34.835 "data_size": 7936 00:17:34.835 } 00:17:34.835 ] 00:17:34.835 }' 00:17:34.835 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:34.835 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:34.835 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:35.093 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:35.093 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:35.093 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@652 -- # local es=0 00:17:35.093 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:35.093 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:17:35.093 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:35.093 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:17:35.093 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:35.093 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:35.093 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:35.093 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:35.094 [2024-11-20 16:05:33.099073] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:35.094 [2024-11-20 16:05:33.099187] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:17:35.094 [2024-11-20 16:05:33.099198] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:17:35.094 request: 00:17:35.094 { 00:17:35.094 "base_bdev": "BaseBdev1", 00:17:35.094 "raid_bdev": "raid_bdev1", 00:17:35.094 "method": "bdev_raid_add_base_bdev", 00:17:35.094 "req_id": 1 00:17:35.094 } 00:17:35.094 Got JSON-RPC error response 00:17:35.094 response: 00:17:35.094 { 00:17:35.094 "code": -22, 00:17:35.094 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:17:35.094 } 00:17:35.094 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:17:35.094 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@655 -- # es=1 00:17:35.094 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:17:35.094 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:17:35.094 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:17:35.094 16:05:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@779 -- # sleep 1 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:36.028 "name": "raid_bdev1", 00:17:36.028 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:36.028 "strip_size_kb": 0, 00:17:36.028 "state": "online", 00:17:36.028 "raid_level": "raid1", 00:17:36.028 "superblock": true, 00:17:36.028 "num_base_bdevs": 2, 00:17:36.028 "num_base_bdevs_discovered": 1, 00:17:36.028 "num_base_bdevs_operational": 1, 00:17:36.028 "base_bdevs_list": [ 00:17:36.028 { 00:17:36.028 "name": null, 00:17:36.028 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:36.028 "is_configured": false, 00:17:36.028 "data_offset": 0, 00:17:36.028 "data_size": 7936 00:17:36.028 }, 00:17:36.028 { 00:17:36.028 "name": "BaseBdev2", 00:17:36.028 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:36.028 "is_configured": true, 00:17:36.028 "data_offset": 256, 00:17:36.028 "data_size": 7936 00:17:36.028 } 00:17:36.028 ] 00:17:36.028 }' 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:36.028 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:36.285 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:36.286 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:36.286 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:36.286 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:36.286 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:36.286 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:36.286 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:36.286 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:36.286 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:36.286 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:36.286 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:36.286 "name": "raid_bdev1", 00:17:36.286 "uuid": "eb93533e-b04f-410a-ab5e-f6c0e057a5a1", 00:17:36.286 "strip_size_kb": 0, 00:17:36.286 "state": "online", 00:17:36.286 "raid_level": "raid1", 00:17:36.286 "superblock": true, 00:17:36.286 "num_base_bdevs": 2, 00:17:36.286 "num_base_bdevs_discovered": 1, 00:17:36.286 "num_base_bdevs_operational": 1, 00:17:36.286 "base_bdevs_list": [ 00:17:36.286 { 00:17:36.286 "name": null, 00:17:36.286 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:36.286 "is_configured": false, 00:17:36.286 "data_offset": 0, 00:17:36.286 "data_size": 7936 00:17:36.286 }, 00:17:36.286 { 00:17:36.286 "name": "BaseBdev2", 00:17:36.286 "uuid": "c82aeff5-1e23-56e7-b785-a28c217fc458", 00:17:36.286 "is_configured": true, 00:17:36.286 "data_offset": 256, 00:17:36.286 "data_size": 7936 00:17:36.286 } 00:17:36.286 ] 00:17:36.286 }' 00:17:36.286 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:36.286 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:36.286 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:36.544 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:36.544 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@784 -- # killprocess 85278 00:17:36.544 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@954 -- # '[' -z 85278 ']' 00:17:36.544 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@958 -- # kill -0 85278 00:17:36.544 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@959 -- # uname 00:17:36.544 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:36.544 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85278 00:17:36.544 killing process with pid 85278 00:17:36.544 Received shutdown signal, test time was about 60.000000 seconds 00:17:36.544 00:17:36.544 Latency(us) 00:17:36.544 [2024-11-20T16:05:34.795Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:36.544 [2024-11-20T16:05:34.795Z] =================================================================================================================== 00:17:36.544 [2024-11-20T16:05:34.795Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:17:36.544 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:36.544 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:36.544 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85278' 00:17:36.544 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@973 -- # kill 85278 00:17:36.544 [2024-11-20 16:05:34.569433] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:36.544 [2024-11-20 16:05:34.569523] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:36.544 16:05:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@978 -- # wait 85278 00:17:36.544 [2024-11-20 16:05:34.569558] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:36.544 [2024-11-20 16:05:34.569566] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:17:36.544 [2024-11-20 16:05:34.725575] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:37.223 16:05:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@786 -- # return 0 00:17:37.223 00:17:37.223 real 0m17.150s 00:17:37.223 user 0m21.778s 00:17:37.223 sys 0m1.880s 00:17:37.223 ************************************ 00:17:37.223 END TEST raid_rebuild_test_sb_md_separate 00:17:37.223 ************************************ 00:17:37.223 16:05:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:37.223 16:05:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:37.223 16:05:35 bdev_raid -- bdev/bdev_raid.sh@1010 -- # base_malloc_params='-m 32 -i' 00:17:37.223 16:05:35 bdev_raid -- bdev/bdev_raid.sh@1011 -- # run_test raid_state_function_test_sb_md_interleaved raid_state_function_test raid1 2 true 00:17:37.223 16:05:35 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:17:37.223 16:05:35 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:37.223 16:05:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:37.223 ************************************ 00:17:37.223 START TEST raid_state_function_test_sb_md_interleaved 00:17:37.223 ************************************ 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:17:37.223 Process raid pid: 85947 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # local strip_size 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@229 -- # raid_pid=85947 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 85947' 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@231 -- # waitforlisten 85947 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 85947 ']' 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:37.223 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:37.223 16:05:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:37.223 [2024-11-20 16:05:35.415583] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:17:37.223 [2024-11-20 16:05:35.415843] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:17:37.485 [2024-11-20 16:05:35.574959] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:37.485 [2024-11-20 16:05:35.677438] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:37.747 [2024-11-20 16:05:35.815686] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:37.747 [2024-11-20 16:05:35.815878] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:38.322 [2024-11-20 16:05:36.335216] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:17:38.322 [2024-11-20 16:05:36.335266] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:17:38.322 [2024-11-20 16:05:36.335276] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:17:38.322 [2024-11-20 16:05:36.335287] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:38.322 "name": "Existed_Raid", 00:17:38.322 "uuid": "c3ba2d8e-633a-44c6-a202-7902bcb23639", 00:17:38.322 "strip_size_kb": 0, 00:17:38.322 "state": "configuring", 00:17:38.322 "raid_level": "raid1", 00:17:38.322 "superblock": true, 00:17:38.322 "num_base_bdevs": 2, 00:17:38.322 "num_base_bdevs_discovered": 0, 00:17:38.322 "num_base_bdevs_operational": 2, 00:17:38.322 "base_bdevs_list": [ 00:17:38.322 { 00:17:38.322 "name": "BaseBdev1", 00:17:38.322 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:38.322 "is_configured": false, 00:17:38.322 "data_offset": 0, 00:17:38.322 "data_size": 0 00:17:38.322 }, 00:17:38.322 { 00:17:38.322 "name": "BaseBdev2", 00:17:38.322 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:38.322 "is_configured": false, 00:17:38.322 "data_offset": 0, 00:17:38.322 "data_size": 0 00:17:38.322 } 00:17:38.322 ] 00:17:38.322 }' 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:38.322 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:38.583 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:17:38.583 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:38.583 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:38.583 [2024-11-20 16:05:36.663233] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:17:38.583 [2024-11-20 16:05:36.663267] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:17:38.583 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:38.583 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:17:38.583 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:38.583 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:38.583 [2024-11-20 16:05:36.671241] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:17:38.583 [2024-11-20 16:05:36.671282] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:17:38.583 [2024-11-20 16:05:36.671291] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:17:38.583 [2024-11-20 16:05:36.671304] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:17:38.583 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:38.583 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1 00:17:38.583 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:38.583 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:38.583 [2024-11-20 16:05:36.705847] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:38.583 BaseBdev1 00:17:38.583 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:38.583 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:17:38.583 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:17:38.583 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:17:38.583 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@905 -- # local i 00:17:38.583 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:38.584 [ 00:17:38.584 { 00:17:38.584 "name": "BaseBdev1", 00:17:38.584 "aliases": [ 00:17:38.584 "83265253-a8cd-420c-9e0f-68b0c4bc286e" 00:17:38.584 ], 00:17:38.584 "product_name": "Malloc disk", 00:17:38.584 "block_size": 4128, 00:17:38.584 "num_blocks": 8192, 00:17:38.584 "uuid": "83265253-a8cd-420c-9e0f-68b0c4bc286e", 00:17:38.584 "md_size": 32, 00:17:38.584 "md_interleave": true, 00:17:38.584 "dif_type": 0, 00:17:38.584 "assigned_rate_limits": { 00:17:38.584 "rw_ios_per_sec": 0, 00:17:38.584 "rw_mbytes_per_sec": 0, 00:17:38.584 "r_mbytes_per_sec": 0, 00:17:38.584 "w_mbytes_per_sec": 0 00:17:38.584 }, 00:17:38.584 "claimed": true, 00:17:38.584 "claim_type": "exclusive_write", 00:17:38.584 "zoned": false, 00:17:38.584 "supported_io_types": { 00:17:38.584 "read": true, 00:17:38.584 "write": true, 00:17:38.584 "unmap": true, 00:17:38.584 "flush": true, 00:17:38.584 "reset": true, 00:17:38.584 "nvme_admin": false, 00:17:38.584 "nvme_io": false, 00:17:38.584 "nvme_io_md": false, 00:17:38.584 "write_zeroes": true, 00:17:38.584 "zcopy": true, 00:17:38.584 "get_zone_info": false, 00:17:38.584 "zone_management": false, 00:17:38.584 "zone_append": false, 00:17:38.584 "compare": false, 00:17:38.584 "compare_and_write": false, 00:17:38.584 "abort": true, 00:17:38.584 "seek_hole": false, 00:17:38.584 "seek_data": false, 00:17:38.584 "copy": true, 00:17:38.584 "nvme_iov_md": false 00:17:38.584 }, 00:17:38.584 "memory_domains": [ 00:17:38.584 { 00:17:38.584 "dma_device_id": "system", 00:17:38.584 "dma_device_type": 1 00:17:38.584 }, 00:17:38.584 { 00:17:38.584 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:38.584 "dma_device_type": 2 00:17:38.584 } 00:17:38.584 ], 00:17:38.584 "driver_specific": {} 00:17:38.584 } 00:17:38.584 ] 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@911 -- # return 0 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:38.584 "name": "Existed_Raid", 00:17:38.584 "uuid": "c39de03a-7b60-4f0b-a8b6-e510ae24719e", 00:17:38.584 "strip_size_kb": 0, 00:17:38.584 "state": "configuring", 00:17:38.584 "raid_level": "raid1", 00:17:38.584 "superblock": true, 00:17:38.584 "num_base_bdevs": 2, 00:17:38.584 "num_base_bdevs_discovered": 1, 00:17:38.584 "num_base_bdevs_operational": 2, 00:17:38.584 "base_bdevs_list": [ 00:17:38.584 { 00:17:38.584 "name": "BaseBdev1", 00:17:38.584 "uuid": "83265253-a8cd-420c-9e0f-68b0c4bc286e", 00:17:38.584 "is_configured": true, 00:17:38.584 "data_offset": 256, 00:17:38.584 "data_size": 7936 00:17:38.584 }, 00:17:38.584 { 00:17:38.584 "name": "BaseBdev2", 00:17:38.584 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:38.584 "is_configured": false, 00:17:38.584 "data_offset": 0, 00:17:38.584 "data_size": 0 00:17:38.584 } 00:17:38.584 ] 00:17:38.584 }' 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:38.584 16:05:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:38.846 [2024-11-20 16:05:37.062040] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:17:38.846 [2024-11-20 16:05:37.062101] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:38.846 [2024-11-20 16:05:37.070073] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:38.846 [2024-11-20 16:05:37.072303] bdev.c:8626:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:17:38.846 [2024-11-20 16:05:37.072349] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:38.846 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:39.106 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:39.106 "name": "Existed_Raid", 00:17:39.106 "uuid": "5fa4e5fe-af67-4e8f-b4ee-363a38c2ebed", 00:17:39.106 "strip_size_kb": 0, 00:17:39.106 "state": "configuring", 00:17:39.106 "raid_level": "raid1", 00:17:39.106 "superblock": true, 00:17:39.106 "num_base_bdevs": 2, 00:17:39.106 "num_base_bdevs_discovered": 1, 00:17:39.106 "num_base_bdevs_operational": 2, 00:17:39.106 "base_bdevs_list": [ 00:17:39.106 { 00:17:39.106 "name": "BaseBdev1", 00:17:39.106 "uuid": "83265253-a8cd-420c-9e0f-68b0c4bc286e", 00:17:39.106 "is_configured": true, 00:17:39.106 "data_offset": 256, 00:17:39.107 "data_size": 7936 00:17:39.107 }, 00:17:39.107 { 00:17:39.107 "name": "BaseBdev2", 00:17:39.107 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:39.107 "is_configured": false, 00:17:39.107 "data_offset": 0, 00:17:39.107 "data_size": 0 00:17:39.107 } 00:17:39.107 ] 00:17:39.107 }' 00:17:39.107 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:39.107 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:39.369 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2 00:17:39.369 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:39.369 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:39.369 [2024-11-20 16:05:37.416950] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:39.369 [2024-11-20 16:05:37.417139] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:17:39.369 [2024-11-20 16:05:37.417153] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:39.369 [2024-11-20 16:05:37.417231] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:17:39.369 [2024-11-20 16:05:37.417299] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:17:39.369 [2024-11-20 16:05:37.417309] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:17:39.369 [2024-11-20 16:05:37.417363] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:39.369 BaseBdev2 00:17:39.369 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:39.369 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:17:39.369 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:17:39.369 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:17:39.369 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@905 -- # local i 00:17:39.369 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:17:39.369 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:39.370 [ 00:17:39.370 { 00:17:39.370 "name": "BaseBdev2", 00:17:39.370 "aliases": [ 00:17:39.370 "ec13a824-be50-4f22-8f8a-608fa5fe61a8" 00:17:39.370 ], 00:17:39.370 "product_name": "Malloc disk", 00:17:39.370 "block_size": 4128, 00:17:39.370 "num_blocks": 8192, 00:17:39.370 "uuid": "ec13a824-be50-4f22-8f8a-608fa5fe61a8", 00:17:39.370 "md_size": 32, 00:17:39.370 "md_interleave": true, 00:17:39.370 "dif_type": 0, 00:17:39.370 "assigned_rate_limits": { 00:17:39.370 "rw_ios_per_sec": 0, 00:17:39.370 "rw_mbytes_per_sec": 0, 00:17:39.370 "r_mbytes_per_sec": 0, 00:17:39.370 "w_mbytes_per_sec": 0 00:17:39.370 }, 00:17:39.370 "claimed": true, 00:17:39.370 "claim_type": "exclusive_write", 00:17:39.370 "zoned": false, 00:17:39.370 "supported_io_types": { 00:17:39.370 "read": true, 00:17:39.370 "write": true, 00:17:39.370 "unmap": true, 00:17:39.370 "flush": true, 00:17:39.370 "reset": true, 00:17:39.370 "nvme_admin": false, 00:17:39.370 "nvme_io": false, 00:17:39.370 "nvme_io_md": false, 00:17:39.370 "write_zeroes": true, 00:17:39.370 "zcopy": true, 00:17:39.370 "get_zone_info": false, 00:17:39.370 "zone_management": false, 00:17:39.370 "zone_append": false, 00:17:39.370 "compare": false, 00:17:39.370 "compare_and_write": false, 00:17:39.370 "abort": true, 00:17:39.370 "seek_hole": false, 00:17:39.370 "seek_data": false, 00:17:39.370 "copy": true, 00:17:39.370 "nvme_iov_md": false 00:17:39.370 }, 00:17:39.370 "memory_domains": [ 00:17:39.370 { 00:17:39.370 "dma_device_id": "system", 00:17:39.370 "dma_device_type": 1 00:17:39.370 }, 00:17:39.370 { 00:17:39.370 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:39.370 "dma_device_type": 2 00:17:39.370 } 00:17:39.370 ], 00:17:39.370 "driver_specific": {} 00:17:39.370 } 00:17:39.370 ] 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@911 -- # return 0 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:39.370 "name": "Existed_Raid", 00:17:39.370 "uuid": "5fa4e5fe-af67-4e8f-b4ee-363a38c2ebed", 00:17:39.370 "strip_size_kb": 0, 00:17:39.370 "state": "online", 00:17:39.370 "raid_level": "raid1", 00:17:39.370 "superblock": true, 00:17:39.370 "num_base_bdevs": 2, 00:17:39.370 "num_base_bdevs_discovered": 2, 00:17:39.370 "num_base_bdevs_operational": 2, 00:17:39.370 "base_bdevs_list": [ 00:17:39.370 { 00:17:39.370 "name": "BaseBdev1", 00:17:39.370 "uuid": "83265253-a8cd-420c-9e0f-68b0c4bc286e", 00:17:39.370 "is_configured": true, 00:17:39.370 "data_offset": 256, 00:17:39.370 "data_size": 7936 00:17:39.370 }, 00:17:39.370 { 00:17:39.370 "name": "BaseBdev2", 00:17:39.370 "uuid": "ec13a824-be50-4f22-8f8a-608fa5fe61a8", 00:17:39.370 "is_configured": true, 00:17:39.370 "data_offset": 256, 00:17:39.370 "data_size": 7936 00:17:39.370 } 00:17:39.370 ] 00:17:39.370 }' 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:39.370 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:39.629 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:17:39.629 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:17:39.630 [2024-11-20 16:05:37.761402] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:17:39.630 "name": "Existed_Raid", 00:17:39.630 "aliases": [ 00:17:39.630 "5fa4e5fe-af67-4e8f-b4ee-363a38c2ebed" 00:17:39.630 ], 00:17:39.630 "product_name": "Raid Volume", 00:17:39.630 "block_size": 4128, 00:17:39.630 "num_blocks": 7936, 00:17:39.630 "uuid": "5fa4e5fe-af67-4e8f-b4ee-363a38c2ebed", 00:17:39.630 "md_size": 32, 00:17:39.630 "md_interleave": true, 00:17:39.630 "dif_type": 0, 00:17:39.630 "assigned_rate_limits": { 00:17:39.630 "rw_ios_per_sec": 0, 00:17:39.630 "rw_mbytes_per_sec": 0, 00:17:39.630 "r_mbytes_per_sec": 0, 00:17:39.630 "w_mbytes_per_sec": 0 00:17:39.630 }, 00:17:39.630 "claimed": false, 00:17:39.630 "zoned": false, 00:17:39.630 "supported_io_types": { 00:17:39.630 "read": true, 00:17:39.630 "write": true, 00:17:39.630 "unmap": false, 00:17:39.630 "flush": false, 00:17:39.630 "reset": true, 00:17:39.630 "nvme_admin": false, 00:17:39.630 "nvme_io": false, 00:17:39.630 "nvme_io_md": false, 00:17:39.630 "write_zeroes": true, 00:17:39.630 "zcopy": false, 00:17:39.630 "get_zone_info": false, 00:17:39.630 "zone_management": false, 00:17:39.630 "zone_append": false, 00:17:39.630 "compare": false, 00:17:39.630 "compare_and_write": false, 00:17:39.630 "abort": false, 00:17:39.630 "seek_hole": false, 00:17:39.630 "seek_data": false, 00:17:39.630 "copy": false, 00:17:39.630 "nvme_iov_md": false 00:17:39.630 }, 00:17:39.630 "memory_domains": [ 00:17:39.630 { 00:17:39.630 "dma_device_id": "system", 00:17:39.630 "dma_device_type": 1 00:17:39.630 }, 00:17:39.630 { 00:17:39.630 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:39.630 "dma_device_type": 2 00:17:39.630 }, 00:17:39.630 { 00:17:39.630 "dma_device_id": "system", 00:17:39.630 "dma_device_type": 1 00:17:39.630 }, 00:17:39.630 { 00:17:39.630 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:39.630 "dma_device_type": 2 00:17:39.630 } 00:17:39.630 ], 00:17:39.630 "driver_specific": { 00:17:39.630 "raid": { 00:17:39.630 "uuid": "5fa4e5fe-af67-4e8f-b4ee-363a38c2ebed", 00:17:39.630 "strip_size_kb": 0, 00:17:39.630 "state": "online", 00:17:39.630 "raid_level": "raid1", 00:17:39.630 "superblock": true, 00:17:39.630 "num_base_bdevs": 2, 00:17:39.630 "num_base_bdevs_discovered": 2, 00:17:39.630 "num_base_bdevs_operational": 2, 00:17:39.630 "base_bdevs_list": [ 00:17:39.630 { 00:17:39.630 "name": "BaseBdev1", 00:17:39.630 "uuid": "83265253-a8cd-420c-9e0f-68b0c4bc286e", 00:17:39.630 "is_configured": true, 00:17:39.630 "data_offset": 256, 00:17:39.630 "data_size": 7936 00:17:39.630 }, 00:17:39.630 { 00:17:39.630 "name": "BaseBdev2", 00:17:39.630 "uuid": "ec13a824-be50-4f22-8f8a-608fa5fe61a8", 00:17:39.630 "is_configured": true, 00:17:39.630 "data_offset": 256, 00:17:39.630 "data_size": 7936 00:17:39.630 } 00:17:39.630 ] 00:17:39.630 } 00:17:39.630 } 00:17:39.630 }' 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:17:39.630 BaseBdev2' 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:39.630 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:39.891 [2024-11-20 16:05:37.921149] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@260 -- # local expected_state 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:39.891 16:05:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:39.891 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:39.891 "name": "Existed_Raid", 00:17:39.891 "uuid": "5fa4e5fe-af67-4e8f-b4ee-363a38c2ebed", 00:17:39.891 "strip_size_kb": 0, 00:17:39.891 "state": "online", 00:17:39.891 "raid_level": "raid1", 00:17:39.891 "superblock": true, 00:17:39.891 "num_base_bdevs": 2, 00:17:39.891 "num_base_bdevs_discovered": 1, 00:17:39.891 "num_base_bdevs_operational": 1, 00:17:39.891 "base_bdevs_list": [ 00:17:39.891 { 00:17:39.891 "name": null, 00:17:39.891 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:39.891 "is_configured": false, 00:17:39.891 "data_offset": 0, 00:17:39.891 "data_size": 7936 00:17:39.891 }, 00:17:39.891 { 00:17:39.891 "name": "BaseBdev2", 00:17:39.891 "uuid": "ec13a824-be50-4f22-8f8a-608fa5fe61a8", 00:17:39.891 "is_configured": true, 00:17:39.891 "data_offset": 256, 00:17:39.891 "data_size": 7936 00:17:39.891 } 00:17:39.891 ] 00:17:39.891 }' 00:17:39.891 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:39.891 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:40.154 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:17:40.154 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:17:40.154 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:40.154 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:40.154 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:40.154 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:17:40.154 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:40.154 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:17:40.154 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:17:40.154 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:17:40.154 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:40.154 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:40.154 [2024-11-20 16:05:38.315967] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:17:40.154 [2024-11-20 16:05:38.316062] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:40.154 [2024-11-20 16:05:38.374417] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:40.154 [2024-11-20 16:05:38.374459] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:40.154 [2024-11-20 16:05:38.374470] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:17:40.154 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:40.154 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:17:40.154 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:17:40.154 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:40.154 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:40.155 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:40.155 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:17:40.155 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:40.417 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:17:40.417 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:17:40.417 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:17:40.417 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@326 -- # killprocess 85947 00:17:40.417 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 85947 ']' 00:17:40.417 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 85947 00:17:40.417 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:17:40.417 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:40.417 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85947 00:17:40.417 killing process with pid 85947 00:17:40.417 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:40.417 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:40.417 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85947' 00:17:40.417 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@973 -- # kill 85947 00:17:40.417 [2024-11-20 16:05:38.440711] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:40.417 16:05:38 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@978 -- # wait 85947 00:17:40.417 [2024-11-20 16:05:38.451164] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:40.990 ************************************ 00:17:40.990 END TEST raid_state_function_test_sb_md_interleaved 00:17:40.990 ************************************ 00:17:40.990 16:05:39 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@328 -- # return 0 00:17:40.990 00:17:40.990 real 0m3.814s 00:17:40.990 user 0m5.522s 00:17:40.990 sys 0m0.607s 00:17:40.990 16:05:39 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:40.990 16:05:39 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:40.990 16:05:39 bdev_raid -- bdev/bdev_raid.sh@1012 -- # run_test raid_superblock_test_md_interleaved raid_superblock_test raid1 2 00:17:40.990 16:05:39 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:17:40.990 16:05:39 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:40.990 16:05:39 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:40.990 ************************************ 00:17:40.990 START TEST raid_superblock_test_md_interleaved 00:17:40.990 ************************************ 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@399 -- # local strip_size 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@412 -- # raid_pid=86179 00:17:40.990 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@413 -- # waitforlisten 86179 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 86179 ']' 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:40.990 16:05:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:41.251 [2024-11-20 16:05:39.291413] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:17:41.251 [2024-11-20 16:05:39.291654] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86179 ] 00:17:41.251 [2024-11-20 16:05:39.452105] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:41.512 [2024-11-20 16:05:39.550997] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:41.512 [2024-11-20 16:05:39.686736] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:41.512 [2024-11-20 16:05:39.686868] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc1 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.085 malloc1 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.085 [2024-11-20 16:05:40.168292] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:17:42.085 [2024-11-20 16:05:40.168346] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:42.085 [2024-11-20 16:05:40.168367] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:17:42.085 [2024-11-20 16:05:40.168376] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:42.085 [2024-11-20 16:05:40.170242] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:42.085 [2024-11-20 16:05:40.170274] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:17:42.085 pt1 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc2 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.085 malloc2 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.085 [2024-11-20 16:05:40.204089] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:42.085 [2024-11-20 16:05:40.204137] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:42.085 [2024-11-20 16:05:40.204157] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:17:42.085 [2024-11-20 16:05:40.204166] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:42.085 [2024-11-20 16:05:40.206012] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:42.085 [2024-11-20 16:05:40.206136] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:42.085 pt2 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.085 [2024-11-20 16:05:40.212120] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:17:42.085 [2024-11-20 16:05:40.213931] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:42.085 [2024-11-20 16:05:40.214097] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:17:42.085 [2024-11-20 16:05:40.214109] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:42.085 [2024-11-20 16:05:40.214178] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:17:42.085 [2024-11-20 16:05:40.214245] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:17:42.085 [2024-11-20 16:05:40.214255] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:17:42.085 [2024-11-20 16:05:40.214317] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.085 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:42.086 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.086 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:42.086 "name": "raid_bdev1", 00:17:42.086 "uuid": "44d49290-d0cb-4aed-b024-4ceb73027e4c", 00:17:42.086 "strip_size_kb": 0, 00:17:42.086 "state": "online", 00:17:42.086 "raid_level": "raid1", 00:17:42.086 "superblock": true, 00:17:42.086 "num_base_bdevs": 2, 00:17:42.086 "num_base_bdevs_discovered": 2, 00:17:42.086 "num_base_bdevs_operational": 2, 00:17:42.086 "base_bdevs_list": [ 00:17:42.086 { 00:17:42.086 "name": "pt1", 00:17:42.086 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:42.086 "is_configured": true, 00:17:42.086 "data_offset": 256, 00:17:42.086 "data_size": 7936 00:17:42.086 }, 00:17:42.086 { 00:17:42.086 "name": "pt2", 00:17:42.086 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:42.086 "is_configured": true, 00:17:42.086 "data_offset": 256, 00:17:42.086 "data_size": 7936 00:17:42.086 } 00:17:42.086 ] 00:17:42.086 }' 00:17:42.086 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:42.086 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.347 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:17:42.347 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:17:42.347 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:17:42.347 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:17:42.347 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:17:42.347 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:17:42.347 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:42.347 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:17:42.347 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.347 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.347 [2024-11-20 16:05:40.564467] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:42.347 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.347 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:17:42.347 "name": "raid_bdev1", 00:17:42.347 "aliases": [ 00:17:42.347 "44d49290-d0cb-4aed-b024-4ceb73027e4c" 00:17:42.347 ], 00:17:42.347 "product_name": "Raid Volume", 00:17:42.347 "block_size": 4128, 00:17:42.348 "num_blocks": 7936, 00:17:42.348 "uuid": "44d49290-d0cb-4aed-b024-4ceb73027e4c", 00:17:42.348 "md_size": 32, 00:17:42.348 "md_interleave": true, 00:17:42.348 "dif_type": 0, 00:17:42.348 "assigned_rate_limits": { 00:17:42.348 "rw_ios_per_sec": 0, 00:17:42.348 "rw_mbytes_per_sec": 0, 00:17:42.348 "r_mbytes_per_sec": 0, 00:17:42.348 "w_mbytes_per_sec": 0 00:17:42.348 }, 00:17:42.348 "claimed": false, 00:17:42.348 "zoned": false, 00:17:42.348 "supported_io_types": { 00:17:42.348 "read": true, 00:17:42.348 "write": true, 00:17:42.348 "unmap": false, 00:17:42.348 "flush": false, 00:17:42.348 "reset": true, 00:17:42.348 "nvme_admin": false, 00:17:42.348 "nvme_io": false, 00:17:42.348 "nvme_io_md": false, 00:17:42.348 "write_zeroes": true, 00:17:42.348 "zcopy": false, 00:17:42.348 "get_zone_info": false, 00:17:42.348 "zone_management": false, 00:17:42.348 "zone_append": false, 00:17:42.348 "compare": false, 00:17:42.348 "compare_and_write": false, 00:17:42.348 "abort": false, 00:17:42.348 "seek_hole": false, 00:17:42.348 "seek_data": false, 00:17:42.348 "copy": false, 00:17:42.348 "nvme_iov_md": false 00:17:42.348 }, 00:17:42.348 "memory_domains": [ 00:17:42.348 { 00:17:42.348 "dma_device_id": "system", 00:17:42.348 "dma_device_type": 1 00:17:42.348 }, 00:17:42.348 { 00:17:42.348 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:42.348 "dma_device_type": 2 00:17:42.348 }, 00:17:42.348 { 00:17:42.348 "dma_device_id": "system", 00:17:42.348 "dma_device_type": 1 00:17:42.348 }, 00:17:42.348 { 00:17:42.348 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:42.348 "dma_device_type": 2 00:17:42.348 } 00:17:42.348 ], 00:17:42.348 "driver_specific": { 00:17:42.348 "raid": { 00:17:42.348 "uuid": "44d49290-d0cb-4aed-b024-4ceb73027e4c", 00:17:42.348 "strip_size_kb": 0, 00:17:42.348 "state": "online", 00:17:42.348 "raid_level": "raid1", 00:17:42.348 "superblock": true, 00:17:42.348 "num_base_bdevs": 2, 00:17:42.348 "num_base_bdevs_discovered": 2, 00:17:42.348 "num_base_bdevs_operational": 2, 00:17:42.348 "base_bdevs_list": [ 00:17:42.348 { 00:17:42.348 "name": "pt1", 00:17:42.348 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:42.348 "is_configured": true, 00:17:42.348 "data_offset": 256, 00:17:42.348 "data_size": 7936 00:17:42.348 }, 00:17:42.348 { 00:17:42.348 "name": "pt2", 00:17:42.348 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:42.348 "is_configured": true, 00:17:42.348 "data_offset": 256, 00:17:42.348 "data_size": 7936 00:17:42.348 } 00:17:42.348 ] 00:17:42.348 } 00:17:42.348 } 00:17:42.348 }' 00:17:42.348 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:17:42.607 pt2' 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.607 [2024-11-20 16:05:40.716472] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=44d49290-d0cb-4aed-b024-4ceb73027e4c 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@436 -- # '[' -z 44d49290-d0cb-4aed-b024-4ceb73027e4c ']' 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.607 [2024-11-20 16:05:40.748188] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:42.607 [2024-11-20 16:05:40.748207] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:42.607 [2024-11-20 16:05:40.748272] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:42.607 [2024-11-20 16:05:40.748325] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:42.607 [2024-11-20 16:05:40.748336] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@652 -- # local es=0 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:17:42.607 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:42.608 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:17:42.608 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:42.608 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:17:42.608 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.608 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.608 [2024-11-20 16:05:40.840238] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:17:42.608 [2024-11-20 16:05:40.842161] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:17:42.608 [2024-11-20 16:05:40.842222] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:17:42.608 [2024-11-20 16:05:40.842271] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:17:42.608 [2024-11-20 16:05:40.842286] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:42.608 [2024-11-20 16:05:40.842296] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:17:42.608 request: 00:17:42.608 { 00:17:42.608 "name": "raid_bdev1", 00:17:42.608 "raid_level": "raid1", 00:17:42.608 "base_bdevs": [ 00:17:42.608 "malloc1", 00:17:42.608 "malloc2" 00:17:42.608 ], 00:17:42.608 "superblock": false, 00:17:42.608 "method": "bdev_raid_create", 00:17:42.608 "req_id": 1 00:17:42.608 } 00:17:42.608 Got JSON-RPC error response 00:17:42.608 response: 00:17:42.608 { 00:17:42.608 "code": -17, 00:17:42.608 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:17:42.608 } 00:17:42.608 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:17:42.608 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@655 -- # es=1 00:17:42.608 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:17:42.608 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:17:42.608 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:17:42.608 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:17:42.608 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:42.608 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.608 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.868 [2024-11-20 16:05:40.884228] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:17:42.868 [2024-11-20 16:05:40.884272] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:42.868 [2024-11-20 16:05:40.884285] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:17:42.868 [2024-11-20 16:05:40.884295] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:42.868 [2024-11-20 16:05:40.886176] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:42.868 [2024-11-20 16:05:40.886207] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:17:42.868 [2024-11-20 16:05:40.886248] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:17:42.868 [2024-11-20 16:05:40.886294] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:17:42.868 pt1 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:42.868 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:42.869 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.869 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.869 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.869 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:42.869 "name": "raid_bdev1", 00:17:42.869 "uuid": "44d49290-d0cb-4aed-b024-4ceb73027e4c", 00:17:42.869 "strip_size_kb": 0, 00:17:42.869 "state": "configuring", 00:17:42.869 "raid_level": "raid1", 00:17:42.869 "superblock": true, 00:17:42.869 "num_base_bdevs": 2, 00:17:42.869 "num_base_bdevs_discovered": 1, 00:17:42.869 "num_base_bdevs_operational": 2, 00:17:42.869 "base_bdevs_list": [ 00:17:42.869 { 00:17:42.869 "name": "pt1", 00:17:42.869 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:42.869 "is_configured": true, 00:17:42.869 "data_offset": 256, 00:17:42.869 "data_size": 7936 00:17:42.869 }, 00:17:42.869 { 00:17:42.869 "name": null, 00:17:42.869 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:42.869 "is_configured": false, 00:17:42.869 "data_offset": 256, 00:17:42.869 "data_size": 7936 00:17:42.869 } 00:17:42.869 ] 00:17:42.869 }' 00:17:42.869 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:42.869 16:05:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.128 [2024-11-20 16:05:41.196314] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:43.128 [2024-11-20 16:05:41.196374] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:43.128 [2024-11-20 16:05:41.196391] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:17:43.128 [2024-11-20 16:05:41.196401] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:43.128 [2024-11-20 16:05:41.196549] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:43.128 [2024-11-20 16:05:41.196565] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:43.128 [2024-11-20 16:05:41.196610] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:17:43.128 [2024-11-20 16:05:41.196630] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:43.128 [2024-11-20 16:05:41.196719] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:17:43.128 [2024-11-20 16:05:41.196731] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:43.128 [2024-11-20 16:05:41.196798] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:17:43.128 [2024-11-20 16:05:41.196858] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:17:43.128 [2024-11-20 16:05:41.196866] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:17:43.128 [2024-11-20 16:05:41.196923] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:43.128 pt2 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.128 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:43.128 "name": "raid_bdev1", 00:17:43.128 "uuid": "44d49290-d0cb-4aed-b024-4ceb73027e4c", 00:17:43.128 "strip_size_kb": 0, 00:17:43.128 "state": "online", 00:17:43.128 "raid_level": "raid1", 00:17:43.128 "superblock": true, 00:17:43.128 "num_base_bdevs": 2, 00:17:43.128 "num_base_bdevs_discovered": 2, 00:17:43.128 "num_base_bdevs_operational": 2, 00:17:43.129 "base_bdevs_list": [ 00:17:43.129 { 00:17:43.129 "name": "pt1", 00:17:43.129 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:43.129 "is_configured": true, 00:17:43.129 "data_offset": 256, 00:17:43.129 "data_size": 7936 00:17:43.129 }, 00:17:43.129 { 00:17:43.129 "name": "pt2", 00:17:43.129 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:43.129 "is_configured": true, 00:17:43.129 "data_offset": 256, 00:17:43.129 "data_size": 7936 00:17:43.129 } 00:17:43.129 ] 00:17:43.129 }' 00:17:43.129 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:43.129 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:17:43.389 [2024-11-20 16:05:41.516681] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:17:43.389 "name": "raid_bdev1", 00:17:43.389 "aliases": [ 00:17:43.389 "44d49290-d0cb-4aed-b024-4ceb73027e4c" 00:17:43.389 ], 00:17:43.389 "product_name": "Raid Volume", 00:17:43.389 "block_size": 4128, 00:17:43.389 "num_blocks": 7936, 00:17:43.389 "uuid": "44d49290-d0cb-4aed-b024-4ceb73027e4c", 00:17:43.389 "md_size": 32, 00:17:43.389 "md_interleave": true, 00:17:43.389 "dif_type": 0, 00:17:43.389 "assigned_rate_limits": { 00:17:43.389 "rw_ios_per_sec": 0, 00:17:43.389 "rw_mbytes_per_sec": 0, 00:17:43.389 "r_mbytes_per_sec": 0, 00:17:43.389 "w_mbytes_per_sec": 0 00:17:43.389 }, 00:17:43.389 "claimed": false, 00:17:43.389 "zoned": false, 00:17:43.389 "supported_io_types": { 00:17:43.389 "read": true, 00:17:43.389 "write": true, 00:17:43.389 "unmap": false, 00:17:43.389 "flush": false, 00:17:43.389 "reset": true, 00:17:43.389 "nvme_admin": false, 00:17:43.389 "nvme_io": false, 00:17:43.389 "nvme_io_md": false, 00:17:43.389 "write_zeroes": true, 00:17:43.389 "zcopy": false, 00:17:43.389 "get_zone_info": false, 00:17:43.389 "zone_management": false, 00:17:43.389 "zone_append": false, 00:17:43.389 "compare": false, 00:17:43.389 "compare_and_write": false, 00:17:43.389 "abort": false, 00:17:43.389 "seek_hole": false, 00:17:43.389 "seek_data": false, 00:17:43.389 "copy": false, 00:17:43.389 "nvme_iov_md": false 00:17:43.389 }, 00:17:43.389 "memory_domains": [ 00:17:43.389 { 00:17:43.389 "dma_device_id": "system", 00:17:43.389 "dma_device_type": 1 00:17:43.389 }, 00:17:43.389 { 00:17:43.389 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:43.389 "dma_device_type": 2 00:17:43.389 }, 00:17:43.389 { 00:17:43.389 "dma_device_id": "system", 00:17:43.389 "dma_device_type": 1 00:17:43.389 }, 00:17:43.389 { 00:17:43.389 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:43.389 "dma_device_type": 2 00:17:43.389 } 00:17:43.389 ], 00:17:43.389 "driver_specific": { 00:17:43.389 "raid": { 00:17:43.389 "uuid": "44d49290-d0cb-4aed-b024-4ceb73027e4c", 00:17:43.389 "strip_size_kb": 0, 00:17:43.389 "state": "online", 00:17:43.389 "raid_level": "raid1", 00:17:43.389 "superblock": true, 00:17:43.389 "num_base_bdevs": 2, 00:17:43.389 "num_base_bdevs_discovered": 2, 00:17:43.389 "num_base_bdevs_operational": 2, 00:17:43.389 "base_bdevs_list": [ 00:17:43.389 { 00:17:43.389 "name": "pt1", 00:17:43.389 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:43.389 "is_configured": true, 00:17:43.389 "data_offset": 256, 00:17:43.389 "data_size": 7936 00:17:43.389 }, 00:17:43.389 { 00:17:43.389 "name": "pt2", 00:17:43.389 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:43.389 "is_configured": true, 00:17:43.389 "data_offset": 256, 00:17:43.389 "data_size": 7936 00:17:43.389 } 00:17:43.389 ] 00:17:43.389 } 00:17:43.389 } 00:17:43.389 }' 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:17:43.389 pt2' 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.389 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:17:43.650 [2024-11-20 16:05:41.676729] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # '[' 44d49290-d0cb-4aed-b024-4ceb73027e4c '!=' 44d49290-d0cb-4aed-b024-4ceb73027e4c ']' 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.650 [2024-11-20 16:05:41.712456] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:43.650 "name": "raid_bdev1", 00:17:43.650 "uuid": "44d49290-d0cb-4aed-b024-4ceb73027e4c", 00:17:43.650 "strip_size_kb": 0, 00:17:43.650 "state": "online", 00:17:43.650 "raid_level": "raid1", 00:17:43.650 "superblock": true, 00:17:43.650 "num_base_bdevs": 2, 00:17:43.650 "num_base_bdevs_discovered": 1, 00:17:43.650 "num_base_bdevs_operational": 1, 00:17:43.650 "base_bdevs_list": [ 00:17:43.650 { 00:17:43.650 "name": null, 00:17:43.650 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:43.650 "is_configured": false, 00:17:43.650 "data_offset": 0, 00:17:43.650 "data_size": 7936 00:17:43.650 }, 00:17:43.650 { 00:17:43.650 "name": "pt2", 00:17:43.650 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:43.650 "is_configured": true, 00:17:43.650 "data_offset": 256, 00:17:43.650 "data_size": 7936 00:17:43.650 } 00:17:43.650 ] 00:17:43.650 }' 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:43.650 16:05:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.912 [2024-11-20 16:05:42.024519] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:43.912 [2024-11-20 16:05:42.024540] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:43.912 [2024-11-20 16:05:42.024605] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:43.912 [2024-11-20 16:05:42.024648] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:43.912 [2024-11-20 16:05:42.024660] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@519 -- # i=1 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.912 [2024-11-20 16:05:42.076530] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:43.912 [2024-11-20 16:05:42.076576] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:43.912 [2024-11-20 16:05:42.076590] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:17:43.912 [2024-11-20 16:05:42.076600] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:43.912 [2024-11-20 16:05:42.078527] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:43.912 [2024-11-20 16:05:42.078561] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:43.912 [2024-11-20 16:05:42.078606] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:17:43.912 [2024-11-20 16:05:42.078646] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:43.912 [2024-11-20 16:05:42.078716] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:17:43.912 [2024-11-20 16:05:42.078729] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:43.912 [2024-11-20 16:05:42.078823] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:17:43.912 [2024-11-20 16:05:42.078880] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:17:43.912 [2024-11-20 16:05:42.078888] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:17:43.912 [2024-11-20 16:05:42.078944] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:43.912 pt2 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:43.912 "name": "raid_bdev1", 00:17:43.912 "uuid": "44d49290-d0cb-4aed-b024-4ceb73027e4c", 00:17:43.912 "strip_size_kb": 0, 00:17:43.912 "state": "online", 00:17:43.912 "raid_level": "raid1", 00:17:43.912 "superblock": true, 00:17:43.912 "num_base_bdevs": 2, 00:17:43.912 "num_base_bdevs_discovered": 1, 00:17:43.912 "num_base_bdevs_operational": 1, 00:17:43.912 "base_bdevs_list": [ 00:17:43.912 { 00:17:43.912 "name": null, 00:17:43.912 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:43.912 "is_configured": false, 00:17:43.912 "data_offset": 256, 00:17:43.912 "data_size": 7936 00:17:43.912 }, 00:17:43.912 { 00:17:43.912 "name": "pt2", 00:17:43.912 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:43.912 "is_configured": true, 00:17:43.912 "data_offset": 256, 00:17:43.912 "data_size": 7936 00:17:43.912 } 00:17:43.912 ] 00:17:43.912 }' 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:43.912 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.173 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:44.173 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.173 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.173 [2024-11-20 16:05:42.384578] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:44.173 [2024-11-20 16:05:42.384599] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:44.173 [2024-11-20 16:05:42.384653] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:44.173 [2024-11-20 16:05:42.384710] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:44.173 [2024-11-20 16:05:42.384720] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:17:44.173 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.173 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:17:44.173 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:44.173 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.173 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.173 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.173 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:17:44.173 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:17:44.173 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:17:44.173 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:17:44.173 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.173 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.431 [2024-11-20 16:05:42.424616] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:17:44.431 [2024-11-20 16:05:42.424756] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:44.431 [2024-11-20 16:05:42.424797] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:17:44.431 [2024-11-20 16:05:42.424852] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:44.431 [2024-11-20 16:05:42.426789] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:44.431 [2024-11-20 16:05:42.426885] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:17:44.431 [2024-11-20 16:05:42.426972] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:17:44.431 [2024-11-20 16:05:42.427030] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:17:44.431 [2024-11-20 16:05:42.427137] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:17:44.431 [2024-11-20 16:05:42.427234] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:44.432 [2024-11-20 16:05:42.427297] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:17:44.432 [2024-11-20 16:05:42.427352] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:44.432 [2024-11-20 16:05:42.427424] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:17:44.432 [2024-11-20 16:05:42.427432] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:44.432 [2024-11-20 16:05:42.427498] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:17:44.432 [2024-11-20 16:05:42.427550] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:17:44.432 [2024-11-20 16:05:42.427560] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:17:44.432 [2024-11-20 16:05:42.427633] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:44.432 pt1 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:44.432 "name": "raid_bdev1", 00:17:44.432 "uuid": "44d49290-d0cb-4aed-b024-4ceb73027e4c", 00:17:44.432 "strip_size_kb": 0, 00:17:44.432 "state": "online", 00:17:44.432 "raid_level": "raid1", 00:17:44.432 "superblock": true, 00:17:44.432 "num_base_bdevs": 2, 00:17:44.432 "num_base_bdevs_discovered": 1, 00:17:44.432 "num_base_bdevs_operational": 1, 00:17:44.432 "base_bdevs_list": [ 00:17:44.432 { 00:17:44.432 "name": null, 00:17:44.432 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:44.432 "is_configured": false, 00:17:44.432 "data_offset": 256, 00:17:44.432 "data_size": 7936 00:17:44.432 }, 00:17:44.432 { 00:17:44.432 "name": "pt2", 00:17:44.432 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:44.432 "is_configured": true, 00:17:44.432 "data_offset": 256, 00:17:44.432 "data_size": 7936 00:17:44.432 } 00:17:44.432 ] 00:17:44.432 }' 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:44.432 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.690 [2024-11-20 16:05:42.772897] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # '[' 44d49290-d0cb-4aed-b024-4ceb73027e4c '!=' 44d49290-d0cb-4aed-b024-4ceb73027e4c ']' 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@563 -- # killprocess 86179 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 86179 ']' 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 86179 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:44.690 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 86179 00:17:44.691 killing process with pid 86179 00:17:44.691 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:44.691 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:44.691 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 86179' 00:17:44.691 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@973 -- # kill 86179 00:17:44.691 [2024-11-20 16:05:42.821706] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:44.691 [2024-11-20 16:05:42.821773] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:44.691 [2024-11-20 16:05:42.821812] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:44.691 16:05:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@978 -- # wait 86179 00:17:44.691 [2024-11-20 16:05:42.821823] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:17:44.691 [2024-11-20 16:05:42.923984] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:45.259 16:05:43 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@565 -- # return 0 00:17:45.259 00:17:45.259 real 0m4.270s 00:17:45.259 user 0m6.557s 00:17:45.259 sys 0m0.705s 00:17:45.259 16:05:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:45.259 ************************************ 00:17:45.259 END TEST raid_superblock_test_md_interleaved 00:17:45.259 ************************************ 00:17:45.259 16:05:43 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:45.518 16:05:43 bdev_raid -- bdev/bdev_raid.sh@1013 -- # run_test raid_rebuild_test_sb_md_interleaved raid_rebuild_test raid1 2 true false false 00:17:45.518 16:05:43 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:17:45.518 16:05:43 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:45.518 16:05:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:45.518 ************************************ 00:17:45.518 START TEST raid_rebuild_test_sb_md_interleaved 00:17:45.518 ************************************ 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false false 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # local verify=false 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # local strip_size 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@577 -- # local create_arg 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@579 -- # local data_offset 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:17:45.518 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@597 -- # raid_pid=86491 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@598 -- # waitforlisten 86491 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 86491 ']' 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:45.518 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:45.519 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:17:45.519 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:45.519 16:05:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:45.519 I/O size of 3145728 is greater than zero copy threshold (65536). 00:17:45.519 Zero copy mechanism will not be used. 00:17:45.519 [2024-11-20 16:05:43.622631] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:17:45.519 [2024-11-20 16:05:43.622772] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86491 ] 00:17:45.775 [2024-11-20 16:05:43.768896] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:45.775 [2024-11-20 16:05:43.851781] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:45.775 [2024-11-20 16:05:43.961700] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:45.775 [2024-11-20 16:05:43.961737] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1_malloc 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:46.342 BaseBdev1_malloc 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:46.342 [2024-11-20 16:05:44.448136] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:17:46.342 [2024-11-20 16:05:44.448193] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:46.342 [2024-11-20 16:05:44.448212] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:17:46.342 [2024-11-20 16:05:44.448221] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:46.342 [2024-11-20 16:05:44.449791] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:46.342 [2024-11-20 16:05:44.449917] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:17:46.342 BaseBdev1 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2_malloc 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:46.342 BaseBdev2_malloc 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:46.342 [2024-11-20 16:05:44.479312] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:17:46.342 [2024-11-20 16:05:44.479358] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:46.342 [2024-11-20 16:05:44.479373] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:17:46.342 [2024-11-20 16:05:44.479382] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:46.342 [2024-11-20 16:05:44.480875] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:46.342 [2024-11-20 16:05:44.480990] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:17:46.342 BaseBdev2 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b spare_malloc 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:46.342 spare_malloc 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:46.342 spare_delay 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:46.342 [2024-11-20 16:05:44.533311] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:46.342 [2024-11-20 16:05:44.533353] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:46.342 [2024-11-20 16:05:44.533367] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:17:46.342 [2024-11-20 16:05:44.533375] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:46.342 [2024-11-20 16:05:44.534873] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:46.342 [2024-11-20 16:05:44.534992] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:46.342 spare 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:46.342 [2024-11-20 16:05:44.541348] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:46.342 [2024-11-20 16:05:44.542895] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:46.342 [2024-11-20 16:05:44.543088] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:17:46.342 [2024-11-20 16:05:44.543158] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:46.342 [2024-11-20 16:05:44.543235] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:17:46.342 [2024-11-20 16:05:44.543361] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:17:46.342 [2024-11-20 16:05:44.543384] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:17:46.342 [2024-11-20 16:05:44.543495] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:46.342 "name": "raid_bdev1", 00:17:46.342 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:46.342 "strip_size_kb": 0, 00:17:46.342 "state": "online", 00:17:46.342 "raid_level": "raid1", 00:17:46.342 "superblock": true, 00:17:46.342 "num_base_bdevs": 2, 00:17:46.342 "num_base_bdevs_discovered": 2, 00:17:46.342 "num_base_bdevs_operational": 2, 00:17:46.342 "base_bdevs_list": [ 00:17:46.342 { 00:17:46.342 "name": "BaseBdev1", 00:17:46.342 "uuid": "c14adc20-6081-52ac-a922-6bad77f1363e", 00:17:46.342 "is_configured": true, 00:17:46.342 "data_offset": 256, 00:17:46.342 "data_size": 7936 00:17:46.342 }, 00:17:46.342 { 00:17:46.342 "name": "BaseBdev2", 00:17:46.342 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:46.342 "is_configured": true, 00:17:46.342 "data_offset": 256, 00:17:46.342 "data_size": 7936 00:17:46.342 } 00:17:46.342 ] 00:17:46.342 }' 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:46.342 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:46.911 [2024-11-20 16:05:44.869651] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@624 -- # '[' false = true ']' 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:46.911 [2024-11-20 16:05:44.933394] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:46.911 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:46.911 "name": "raid_bdev1", 00:17:46.911 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:46.911 "strip_size_kb": 0, 00:17:46.911 "state": "online", 00:17:46.911 "raid_level": "raid1", 00:17:46.911 "superblock": true, 00:17:46.911 "num_base_bdevs": 2, 00:17:46.911 "num_base_bdevs_discovered": 1, 00:17:46.911 "num_base_bdevs_operational": 1, 00:17:46.912 "base_bdevs_list": [ 00:17:46.912 { 00:17:46.912 "name": null, 00:17:46.912 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:46.912 "is_configured": false, 00:17:46.912 "data_offset": 0, 00:17:46.912 "data_size": 7936 00:17:46.912 }, 00:17:46.912 { 00:17:46.912 "name": "BaseBdev2", 00:17:46.912 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:46.912 "is_configured": true, 00:17:46.912 "data_offset": 256, 00:17:46.912 "data_size": 7936 00:17:46.912 } 00:17:46.912 ] 00:17:46.912 }' 00:17:46.912 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:46.912 16:05:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:47.169 16:05:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:47.169 16:05:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:47.169 16:05:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:47.170 [2024-11-20 16:05:45.265491] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:47.170 [2024-11-20 16:05:45.274817] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:17:47.170 16:05:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:47.170 16:05:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@647 -- # sleep 1 00:17:47.170 [2024-11-20 16:05:45.276358] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:48.105 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:48.105 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:48.105 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:48.105 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:48.105 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:48.105 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:48.105 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.105 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:48.105 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.105 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.105 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:48.105 "name": "raid_bdev1", 00:17:48.105 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:48.105 "strip_size_kb": 0, 00:17:48.105 "state": "online", 00:17:48.105 "raid_level": "raid1", 00:17:48.105 "superblock": true, 00:17:48.105 "num_base_bdevs": 2, 00:17:48.105 "num_base_bdevs_discovered": 2, 00:17:48.105 "num_base_bdevs_operational": 2, 00:17:48.105 "process": { 00:17:48.105 "type": "rebuild", 00:17:48.105 "target": "spare", 00:17:48.105 "progress": { 00:17:48.105 "blocks": 2560, 00:17:48.105 "percent": 32 00:17:48.105 } 00:17:48.105 }, 00:17:48.105 "base_bdevs_list": [ 00:17:48.105 { 00:17:48.105 "name": "spare", 00:17:48.105 "uuid": "48b8c1c1-6011-5f48-b8e9-cf9d7881dc5a", 00:17:48.105 "is_configured": true, 00:17:48.105 "data_offset": 256, 00:17:48.105 "data_size": 7936 00:17:48.105 }, 00:17:48.105 { 00:17:48.105 "name": "BaseBdev2", 00:17:48.105 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:48.105 "is_configured": true, 00:17:48.105 "data_offset": 256, 00:17:48.105 "data_size": 7936 00:17:48.105 } 00:17:48.105 ] 00:17:48.105 }' 00:17:48.105 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:48.105 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:48.105 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.364 [2024-11-20 16:05:46.378520] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:48.364 [2024-11-20 16:05:46.381135] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:48.364 [2024-11-20 16:05:46.381183] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:48.364 [2024-11-20 16:05:46.381195] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:48.364 [2024-11-20 16:05:46.381204] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:48.364 "name": "raid_bdev1", 00:17:48.364 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:48.364 "strip_size_kb": 0, 00:17:48.364 "state": "online", 00:17:48.364 "raid_level": "raid1", 00:17:48.364 "superblock": true, 00:17:48.364 "num_base_bdevs": 2, 00:17:48.364 "num_base_bdevs_discovered": 1, 00:17:48.364 "num_base_bdevs_operational": 1, 00:17:48.364 "base_bdevs_list": [ 00:17:48.364 { 00:17:48.364 "name": null, 00:17:48.364 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:48.364 "is_configured": false, 00:17:48.364 "data_offset": 0, 00:17:48.364 "data_size": 7936 00:17:48.364 }, 00:17:48.364 { 00:17:48.364 "name": "BaseBdev2", 00:17:48.364 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:48.364 "is_configured": true, 00:17:48.364 "data_offset": 256, 00:17:48.364 "data_size": 7936 00:17:48.364 } 00:17:48.364 ] 00:17:48.364 }' 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:48.364 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:48.623 "name": "raid_bdev1", 00:17:48.623 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:48.623 "strip_size_kb": 0, 00:17:48.623 "state": "online", 00:17:48.623 "raid_level": "raid1", 00:17:48.623 "superblock": true, 00:17:48.623 "num_base_bdevs": 2, 00:17:48.623 "num_base_bdevs_discovered": 1, 00:17:48.623 "num_base_bdevs_operational": 1, 00:17:48.623 "base_bdevs_list": [ 00:17:48.623 { 00:17:48.623 "name": null, 00:17:48.623 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:48.623 "is_configured": false, 00:17:48.623 "data_offset": 0, 00:17:48.623 "data_size": 7936 00:17:48.623 }, 00:17:48.623 { 00:17:48.623 "name": "BaseBdev2", 00:17:48.623 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:48.623 "is_configured": true, 00:17:48.623 "data_offset": 256, 00:17:48.623 "data_size": 7936 00:17:48.623 } 00:17:48.623 ] 00:17:48.623 }' 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.623 [2024-11-20 16:05:46.811756] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:48.623 [2024-11-20 16:05:46.821059] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.623 16:05:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@663 -- # sleep 1 00:17:48.623 [2024-11-20 16:05:46.822624] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:49.994 "name": "raid_bdev1", 00:17:49.994 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:49.994 "strip_size_kb": 0, 00:17:49.994 "state": "online", 00:17:49.994 "raid_level": "raid1", 00:17:49.994 "superblock": true, 00:17:49.994 "num_base_bdevs": 2, 00:17:49.994 "num_base_bdevs_discovered": 2, 00:17:49.994 "num_base_bdevs_operational": 2, 00:17:49.994 "process": { 00:17:49.994 "type": "rebuild", 00:17:49.994 "target": "spare", 00:17:49.994 "progress": { 00:17:49.994 "blocks": 2560, 00:17:49.994 "percent": 32 00:17:49.994 } 00:17:49.994 }, 00:17:49.994 "base_bdevs_list": [ 00:17:49.994 { 00:17:49.994 "name": "spare", 00:17:49.994 "uuid": "48b8c1c1-6011-5f48-b8e9-cf9d7881dc5a", 00:17:49.994 "is_configured": true, 00:17:49.994 "data_offset": 256, 00:17:49.994 "data_size": 7936 00:17:49.994 }, 00:17:49.994 { 00:17:49.994 "name": "BaseBdev2", 00:17:49.994 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:49.994 "is_configured": true, 00:17:49.994 "data_offset": 256, 00:17:49.994 "data_size": 7936 00:17:49.994 } 00:17:49.994 ] 00:17:49.994 }' 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:17:49.994 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@706 -- # local timeout=601 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:49.994 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:49.994 "name": "raid_bdev1", 00:17:49.994 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:49.994 "strip_size_kb": 0, 00:17:49.994 "state": "online", 00:17:49.994 "raid_level": "raid1", 00:17:49.994 "superblock": true, 00:17:49.994 "num_base_bdevs": 2, 00:17:49.994 "num_base_bdevs_discovered": 2, 00:17:49.994 "num_base_bdevs_operational": 2, 00:17:49.994 "process": { 00:17:49.994 "type": "rebuild", 00:17:49.994 "target": "spare", 00:17:49.994 "progress": { 00:17:49.994 "blocks": 2560, 00:17:49.994 "percent": 32 00:17:49.994 } 00:17:49.994 }, 00:17:49.994 "base_bdevs_list": [ 00:17:49.994 { 00:17:49.994 "name": "spare", 00:17:49.994 "uuid": "48b8c1c1-6011-5f48-b8e9-cf9d7881dc5a", 00:17:49.994 "is_configured": true, 00:17:49.994 "data_offset": 256, 00:17:49.994 "data_size": 7936 00:17:49.995 }, 00:17:49.995 { 00:17:49.995 "name": "BaseBdev2", 00:17:49.995 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:49.995 "is_configured": true, 00:17:49.995 "data_offset": 256, 00:17:49.995 "data_size": 7936 00:17:49.995 } 00:17:49.995 ] 00:17:49.995 }' 00:17:49.995 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:49.995 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:49.995 16:05:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:49.995 16:05:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:49.995 16:05:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:17:50.933 16:05:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:50.933 16:05:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:50.933 16:05:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:50.933 16:05:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:50.933 16:05:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:50.933 16:05:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:50.933 16:05:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:50.933 16:05:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:50.933 16:05:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:50.933 16:05:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:50.933 16:05:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:50.933 16:05:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:50.933 "name": "raid_bdev1", 00:17:50.933 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:50.933 "strip_size_kb": 0, 00:17:50.933 "state": "online", 00:17:50.933 "raid_level": "raid1", 00:17:50.933 "superblock": true, 00:17:50.933 "num_base_bdevs": 2, 00:17:50.933 "num_base_bdevs_discovered": 2, 00:17:50.933 "num_base_bdevs_operational": 2, 00:17:50.933 "process": { 00:17:50.933 "type": "rebuild", 00:17:50.933 "target": "spare", 00:17:50.933 "progress": { 00:17:50.933 "blocks": 5376, 00:17:50.933 "percent": 67 00:17:50.933 } 00:17:50.933 }, 00:17:50.933 "base_bdevs_list": [ 00:17:50.933 { 00:17:50.933 "name": "spare", 00:17:50.933 "uuid": "48b8c1c1-6011-5f48-b8e9-cf9d7881dc5a", 00:17:50.933 "is_configured": true, 00:17:50.933 "data_offset": 256, 00:17:50.933 "data_size": 7936 00:17:50.933 }, 00:17:50.933 { 00:17:50.933 "name": "BaseBdev2", 00:17:50.933 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:50.933 "is_configured": true, 00:17:50.933 "data_offset": 256, 00:17:50.933 "data_size": 7936 00:17:50.933 } 00:17:50.933 ] 00:17:50.933 }' 00:17:50.933 16:05:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:50.933 16:05:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:50.933 16:05:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:50.933 16:05:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:50.933 16:05:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:17:51.866 [2024-11-20 16:05:49.936375] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:17:51.866 [2024-11-20 16:05:49.936445] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:17:51.866 [2024-11-20 16:05:49.936544] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:52.125 "name": "raid_bdev1", 00:17:52.125 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:52.125 "strip_size_kb": 0, 00:17:52.125 "state": "online", 00:17:52.125 "raid_level": "raid1", 00:17:52.125 "superblock": true, 00:17:52.125 "num_base_bdevs": 2, 00:17:52.125 "num_base_bdevs_discovered": 2, 00:17:52.125 "num_base_bdevs_operational": 2, 00:17:52.125 "base_bdevs_list": [ 00:17:52.125 { 00:17:52.125 "name": "spare", 00:17:52.125 "uuid": "48b8c1c1-6011-5f48-b8e9-cf9d7881dc5a", 00:17:52.125 "is_configured": true, 00:17:52.125 "data_offset": 256, 00:17:52.125 "data_size": 7936 00:17:52.125 }, 00:17:52.125 { 00:17:52.125 "name": "BaseBdev2", 00:17:52.125 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:52.125 "is_configured": true, 00:17:52.125 "data_offset": 256, 00:17:52.125 "data_size": 7936 00:17:52.125 } 00:17:52.125 ] 00:17:52.125 }' 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@709 -- # break 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:52.125 "name": "raid_bdev1", 00:17:52.125 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:52.125 "strip_size_kb": 0, 00:17:52.125 "state": "online", 00:17:52.125 "raid_level": "raid1", 00:17:52.125 "superblock": true, 00:17:52.125 "num_base_bdevs": 2, 00:17:52.125 "num_base_bdevs_discovered": 2, 00:17:52.125 "num_base_bdevs_operational": 2, 00:17:52.125 "base_bdevs_list": [ 00:17:52.125 { 00:17:52.125 "name": "spare", 00:17:52.125 "uuid": "48b8c1c1-6011-5f48-b8e9-cf9d7881dc5a", 00:17:52.125 "is_configured": true, 00:17:52.125 "data_offset": 256, 00:17:52.125 "data_size": 7936 00:17:52.125 }, 00:17:52.125 { 00:17:52.125 "name": "BaseBdev2", 00:17:52.125 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:52.125 "is_configured": true, 00:17:52.125 "data_offset": 256, 00:17:52.125 "data_size": 7936 00:17:52.125 } 00:17:52.125 ] 00:17:52.125 }' 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.125 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.383 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:52.383 "name": "raid_bdev1", 00:17:52.383 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:52.383 "strip_size_kb": 0, 00:17:52.383 "state": "online", 00:17:52.383 "raid_level": "raid1", 00:17:52.383 "superblock": true, 00:17:52.383 "num_base_bdevs": 2, 00:17:52.383 "num_base_bdevs_discovered": 2, 00:17:52.383 "num_base_bdevs_operational": 2, 00:17:52.383 "base_bdevs_list": [ 00:17:52.383 { 00:17:52.383 "name": "spare", 00:17:52.383 "uuid": "48b8c1c1-6011-5f48-b8e9-cf9d7881dc5a", 00:17:52.383 "is_configured": true, 00:17:52.383 "data_offset": 256, 00:17:52.383 "data_size": 7936 00:17:52.383 }, 00:17:52.383 { 00:17:52.383 "name": "BaseBdev2", 00:17:52.383 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:52.383 "is_configured": true, 00:17:52.383 "data_offset": 256, 00:17:52.383 "data_size": 7936 00:17:52.383 } 00:17:52.383 ] 00:17:52.383 }' 00:17:52.383 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:52.383 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.642 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:52.642 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.642 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.642 [2024-11-20 16:05:50.654862] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:52.642 [2024-11-20 16:05:50.654989] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:52.642 [2024-11-20 16:05:50.655107] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:52.642 [2024-11-20 16:05:50.655213] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:52.642 [2024-11-20 16:05:50.655274] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:17:52.642 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.642 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:52.642 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.642 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.642 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # jq length 00:17:52.642 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.642 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:17:52.642 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:17:52.642 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:17:52.642 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:17:52.642 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.642 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.642 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.642 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.643 [2024-11-20 16:05:50.714854] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:52.643 [2024-11-20 16:05:50.714896] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:52.643 [2024-11-20 16:05:50.714912] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:17:52.643 [2024-11-20 16:05:50.714920] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:52.643 [2024-11-20 16:05:50.716554] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:52.643 [2024-11-20 16:05:50.716583] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:52.643 [2024-11-20 16:05:50.716629] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:17:52.643 [2024-11-20 16:05:50.716680] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:52.643 [2024-11-20 16:05:50.716765] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:52.643 spare 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.643 [2024-11-20 16:05:50.816840] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:17:52.643 [2024-11-20 16:05:50.816873] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:52.643 [2024-11-20 16:05:50.816969] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:17:52.643 [2024-11-20 16:05:50.817048] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:17:52.643 [2024-11-20 16:05:50.817058] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:17:52.643 [2024-11-20 16:05:50.817135] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:52.643 "name": "raid_bdev1", 00:17:52.643 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:52.643 "strip_size_kb": 0, 00:17:52.643 "state": "online", 00:17:52.643 "raid_level": "raid1", 00:17:52.643 "superblock": true, 00:17:52.643 "num_base_bdevs": 2, 00:17:52.643 "num_base_bdevs_discovered": 2, 00:17:52.643 "num_base_bdevs_operational": 2, 00:17:52.643 "base_bdevs_list": [ 00:17:52.643 { 00:17:52.643 "name": "spare", 00:17:52.643 "uuid": "48b8c1c1-6011-5f48-b8e9-cf9d7881dc5a", 00:17:52.643 "is_configured": true, 00:17:52.643 "data_offset": 256, 00:17:52.643 "data_size": 7936 00:17:52.643 }, 00:17:52.643 { 00:17:52.643 "name": "BaseBdev2", 00:17:52.643 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:52.643 "is_configured": true, 00:17:52.643 "data_offset": 256, 00:17:52.643 "data_size": 7936 00:17:52.643 } 00:17:52.643 ] 00:17:52.643 }' 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:52.643 16:05:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.901 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:52.901 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:52.901 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:52.901 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:52.901 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:52.901 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:52.901 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:52.901 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.901 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.901 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:53.158 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:53.158 "name": "raid_bdev1", 00:17:53.158 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:53.158 "strip_size_kb": 0, 00:17:53.158 "state": "online", 00:17:53.158 "raid_level": "raid1", 00:17:53.158 "superblock": true, 00:17:53.158 "num_base_bdevs": 2, 00:17:53.158 "num_base_bdevs_discovered": 2, 00:17:53.158 "num_base_bdevs_operational": 2, 00:17:53.158 "base_bdevs_list": [ 00:17:53.158 { 00:17:53.158 "name": "spare", 00:17:53.158 "uuid": "48b8c1c1-6011-5f48-b8e9-cf9d7881dc5a", 00:17:53.158 "is_configured": true, 00:17:53.158 "data_offset": 256, 00:17:53.158 "data_size": 7936 00:17:53.158 }, 00:17:53.158 { 00:17:53.158 "name": "BaseBdev2", 00:17:53.158 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:53.158 "is_configured": true, 00:17:53.158 "data_offset": 256, 00:17:53.158 "data_size": 7936 00:17:53.158 } 00:17:53.158 ] 00:17:53.158 }' 00:17:53.158 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:53.159 [2024-11-20 16:05:51.263026] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:53.159 "name": "raid_bdev1", 00:17:53.159 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:53.159 "strip_size_kb": 0, 00:17:53.159 "state": "online", 00:17:53.159 "raid_level": "raid1", 00:17:53.159 "superblock": true, 00:17:53.159 "num_base_bdevs": 2, 00:17:53.159 "num_base_bdevs_discovered": 1, 00:17:53.159 "num_base_bdevs_operational": 1, 00:17:53.159 "base_bdevs_list": [ 00:17:53.159 { 00:17:53.159 "name": null, 00:17:53.159 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:53.159 "is_configured": false, 00:17:53.159 "data_offset": 0, 00:17:53.159 "data_size": 7936 00:17:53.159 }, 00:17:53.159 { 00:17:53.159 "name": "BaseBdev2", 00:17:53.159 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:53.159 "is_configured": true, 00:17:53.159 "data_offset": 256, 00:17:53.159 "data_size": 7936 00:17:53.159 } 00:17:53.159 ] 00:17:53.159 }' 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:53.159 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:53.417 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:53.417 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:53.417 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:53.417 [2024-11-20 16:05:51.575091] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:53.417 [2024-11-20 16:05:51.575242] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:17:53.417 [2024-11-20 16:05:51.575255] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:17:53.417 [2024-11-20 16:05:51.575289] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:53.417 [2024-11-20 16:05:51.584056] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:17:53.417 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:53.417 16:05:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@757 -- # sleep 1 00:17:53.417 [2024-11-20 16:05:51.585659] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:54.349 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:54.349 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:54.349 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:54.349 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:54.349 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:54.349 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:54.349 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:54.349 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:54.349 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:54.607 "name": "raid_bdev1", 00:17:54.607 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:54.607 "strip_size_kb": 0, 00:17:54.607 "state": "online", 00:17:54.607 "raid_level": "raid1", 00:17:54.607 "superblock": true, 00:17:54.607 "num_base_bdevs": 2, 00:17:54.607 "num_base_bdevs_discovered": 2, 00:17:54.607 "num_base_bdevs_operational": 2, 00:17:54.607 "process": { 00:17:54.607 "type": "rebuild", 00:17:54.607 "target": "spare", 00:17:54.607 "progress": { 00:17:54.607 "blocks": 2560, 00:17:54.607 "percent": 32 00:17:54.607 } 00:17:54.607 }, 00:17:54.607 "base_bdevs_list": [ 00:17:54.607 { 00:17:54.607 "name": "spare", 00:17:54.607 "uuid": "48b8c1c1-6011-5f48-b8e9-cf9d7881dc5a", 00:17:54.607 "is_configured": true, 00:17:54.607 "data_offset": 256, 00:17:54.607 "data_size": 7936 00:17:54.607 }, 00:17:54.607 { 00:17:54.607 "name": "BaseBdev2", 00:17:54.607 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:54.607 "is_configured": true, 00:17:54.607 "data_offset": 256, 00:17:54.607 "data_size": 7936 00:17:54.607 } 00:17:54.607 ] 00:17:54.607 }' 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:54.607 [2024-11-20 16:05:52.695894] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:54.607 [2024-11-20 16:05:52.791057] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:54.607 [2024-11-20 16:05:52.791124] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:54.607 [2024-11-20 16:05:52.791136] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:54.607 [2024-11-20 16:05:52.791147] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:54.607 "name": "raid_bdev1", 00:17:54.607 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:54.607 "strip_size_kb": 0, 00:17:54.607 "state": "online", 00:17:54.607 "raid_level": "raid1", 00:17:54.607 "superblock": true, 00:17:54.607 "num_base_bdevs": 2, 00:17:54.607 "num_base_bdevs_discovered": 1, 00:17:54.607 "num_base_bdevs_operational": 1, 00:17:54.607 "base_bdevs_list": [ 00:17:54.607 { 00:17:54.607 "name": null, 00:17:54.607 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:54.607 "is_configured": false, 00:17:54.607 "data_offset": 0, 00:17:54.607 "data_size": 7936 00:17:54.607 }, 00:17:54.607 { 00:17:54.607 "name": "BaseBdev2", 00:17:54.607 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:54.607 "is_configured": true, 00:17:54.607 "data_offset": 256, 00:17:54.607 "data_size": 7936 00:17:54.607 } 00:17:54.607 ] 00:17:54.607 }' 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:54.607 16:05:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:55.172 16:05:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:55.172 16:05:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:55.172 16:05:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:55.172 [2024-11-20 16:05:53.129726] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:55.172 [2024-11-20 16:05:53.129781] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:55.172 [2024-11-20 16:05:53.129802] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:17:55.172 [2024-11-20 16:05:53.129812] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:55.172 [2024-11-20 16:05:53.129965] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:55.172 [2024-11-20 16:05:53.129977] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:55.172 [2024-11-20 16:05:53.130020] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:17:55.172 [2024-11-20 16:05:53.130032] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:17:55.172 [2024-11-20 16:05:53.130040] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:17:55.172 [2024-11-20 16:05:53.130057] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:55.172 [2024-11-20 16:05:53.139006] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:17:55.172 spare 00:17:55.172 16:05:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:55.172 16:05:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@764 -- # sleep 1 00:17:55.172 [2024-11-20 16:05:53.140709] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:56.105 "name": "raid_bdev1", 00:17:56.105 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:56.105 "strip_size_kb": 0, 00:17:56.105 "state": "online", 00:17:56.105 "raid_level": "raid1", 00:17:56.105 "superblock": true, 00:17:56.105 "num_base_bdevs": 2, 00:17:56.105 "num_base_bdevs_discovered": 2, 00:17:56.105 "num_base_bdevs_operational": 2, 00:17:56.105 "process": { 00:17:56.105 "type": "rebuild", 00:17:56.105 "target": "spare", 00:17:56.105 "progress": { 00:17:56.105 "blocks": 2560, 00:17:56.105 "percent": 32 00:17:56.105 } 00:17:56.105 }, 00:17:56.105 "base_bdevs_list": [ 00:17:56.105 { 00:17:56.105 "name": "spare", 00:17:56.105 "uuid": "48b8c1c1-6011-5f48-b8e9-cf9d7881dc5a", 00:17:56.105 "is_configured": true, 00:17:56.105 "data_offset": 256, 00:17:56.105 "data_size": 7936 00:17:56.105 }, 00:17:56.105 { 00:17:56.105 "name": "BaseBdev2", 00:17:56.105 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:56.105 "is_configured": true, 00:17:56.105 "data_offset": 256, 00:17:56.105 "data_size": 7936 00:17:56.105 } 00:17:56.105 ] 00:17:56.105 }' 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:56.105 [2024-11-20 16:05:54.242888] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:56.105 [2024-11-20 16:05:54.245601] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:56.105 [2024-11-20 16:05:54.245809] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:56.105 [2024-11-20 16:05:54.245837] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:56.105 [2024-11-20 16:05:54.245849] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:56.105 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:56.106 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:56.106 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:56.106 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:56.106 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:56.106 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:56.106 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:56.106 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:56.106 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:56.106 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:56.106 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:56.106 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:56.106 "name": "raid_bdev1", 00:17:56.106 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:56.106 "strip_size_kb": 0, 00:17:56.106 "state": "online", 00:17:56.106 "raid_level": "raid1", 00:17:56.106 "superblock": true, 00:17:56.106 "num_base_bdevs": 2, 00:17:56.106 "num_base_bdevs_discovered": 1, 00:17:56.106 "num_base_bdevs_operational": 1, 00:17:56.106 "base_bdevs_list": [ 00:17:56.106 { 00:17:56.106 "name": null, 00:17:56.106 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:56.106 "is_configured": false, 00:17:56.106 "data_offset": 0, 00:17:56.106 "data_size": 7936 00:17:56.106 }, 00:17:56.106 { 00:17:56.106 "name": "BaseBdev2", 00:17:56.106 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:56.106 "is_configured": true, 00:17:56.106 "data_offset": 256, 00:17:56.106 "data_size": 7936 00:17:56.106 } 00:17:56.106 ] 00:17:56.106 }' 00:17:56.106 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:56.106 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:56.401 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:56.401 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:56.401 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:56.401 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:56.401 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:56.401 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:56.401 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:56.401 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:56.401 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:56.401 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:56.401 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:56.401 "name": "raid_bdev1", 00:17:56.401 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:56.401 "strip_size_kb": 0, 00:17:56.401 "state": "online", 00:17:56.401 "raid_level": "raid1", 00:17:56.401 "superblock": true, 00:17:56.401 "num_base_bdevs": 2, 00:17:56.401 "num_base_bdevs_discovered": 1, 00:17:56.401 "num_base_bdevs_operational": 1, 00:17:56.401 "base_bdevs_list": [ 00:17:56.401 { 00:17:56.401 "name": null, 00:17:56.401 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:56.401 "is_configured": false, 00:17:56.401 "data_offset": 0, 00:17:56.401 "data_size": 7936 00:17:56.401 }, 00:17:56.401 { 00:17:56.401 "name": "BaseBdev2", 00:17:56.401 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:56.401 "is_configured": true, 00:17:56.401 "data_offset": 256, 00:17:56.401 "data_size": 7936 00:17:56.401 } 00:17:56.401 ] 00:17:56.401 }' 00:17:56.401 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:56.401 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:56.401 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:56.660 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:56.660 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:17:56.660 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:56.660 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:56.660 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:56.660 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:17:56.660 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:56.660 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:56.660 [2024-11-20 16:05:54.688446] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:17:56.660 [2024-11-20 16:05:54.688495] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:56.660 [2024-11-20 16:05:54.688512] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:17:56.660 [2024-11-20 16:05:54.688520] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:56.660 [2024-11-20 16:05:54.688658] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:56.660 [2024-11-20 16:05:54.688684] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:17:56.660 [2024-11-20 16:05:54.688735] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:17:56.660 [2024-11-20 16:05:54.688746] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:17:56.660 [2024-11-20 16:05:54.688754] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:17:56.660 [2024-11-20 16:05:54.688762] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:17:56.660 BaseBdev1 00:17:56.660 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:56.660 16:05:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@775 -- # sleep 1 00:17:57.592 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:57.592 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:57.592 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:57.592 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:57.592 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:57.592 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:57.592 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:57.592 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:57.592 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:57.592 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:57.592 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:57.592 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:57.592 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:57.592 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:57.592 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:57.592 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:57.592 "name": "raid_bdev1", 00:17:57.592 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:57.592 "strip_size_kb": 0, 00:17:57.592 "state": "online", 00:17:57.592 "raid_level": "raid1", 00:17:57.592 "superblock": true, 00:17:57.592 "num_base_bdevs": 2, 00:17:57.592 "num_base_bdevs_discovered": 1, 00:17:57.592 "num_base_bdevs_operational": 1, 00:17:57.592 "base_bdevs_list": [ 00:17:57.592 { 00:17:57.592 "name": null, 00:17:57.592 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:57.592 "is_configured": false, 00:17:57.592 "data_offset": 0, 00:17:57.592 "data_size": 7936 00:17:57.592 }, 00:17:57.592 { 00:17:57.592 "name": "BaseBdev2", 00:17:57.592 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:57.592 "is_configured": true, 00:17:57.592 "data_offset": 256, 00:17:57.593 "data_size": 7936 00:17:57.593 } 00:17:57.593 ] 00:17:57.593 }' 00:17:57.593 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:57.593 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:57.850 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:57.850 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:57.850 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:57.850 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:57.850 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:57.850 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:57.850 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:57.850 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:57.850 16:05:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:57.850 "name": "raid_bdev1", 00:17:57.850 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:57.850 "strip_size_kb": 0, 00:17:57.850 "state": "online", 00:17:57.850 "raid_level": "raid1", 00:17:57.850 "superblock": true, 00:17:57.850 "num_base_bdevs": 2, 00:17:57.850 "num_base_bdevs_discovered": 1, 00:17:57.850 "num_base_bdevs_operational": 1, 00:17:57.850 "base_bdevs_list": [ 00:17:57.850 { 00:17:57.850 "name": null, 00:17:57.850 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:57.850 "is_configured": false, 00:17:57.850 "data_offset": 0, 00:17:57.850 "data_size": 7936 00:17:57.850 }, 00:17:57.850 { 00:17:57.850 "name": "BaseBdev2", 00:17:57.850 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:57.850 "is_configured": true, 00:17:57.850 "data_offset": 256, 00:17:57.850 "data_size": 7936 00:17:57.850 } 00:17:57.850 ] 00:17:57.850 }' 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@652 -- # local es=0 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:57.850 [2024-11-20 16:05:56.084744] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:57.850 [2024-11-20 16:05:56.084865] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:17:57.850 [2024-11-20 16:05:56.084879] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:17:57.850 request: 00:17:57.850 { 00:17:57.850 "base_bdev": "BaseBdev1", 00:17:57.850 "raid_bdev": "raid_bdev1", 00:17:57.850 "method": "bdev_raid_add_base_bdev", 00:17:57.850 "req_id": 1 00:17:57.850 } 00:17:57.850 Got JSON-RPC error response 00:17:57.850 response: 00:17:57.850 { 00:17:57.850 "code": -22, 00:17:57.850 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:17:57.850 } 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@655 -- # es=1 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:17:57.850 16:05:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@779 -- # sleep 1 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:59.224 "name": "raid_bdev1", 00:17:59.224 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:59.224 "strip_size_kb": 0, 00:17:59.224 "state": "online", 00:17:59.224 "raid_level": "raid1", 00:17:59.224 "superblock": true, 00:17:59.224 "num_base_bdevs": 2, 00:17:59.224 "num_base_bdevs_discovered": 1, 00:17:59.224 "num_base_bdevs_operational": 1, 00:17:59.224 "base_bdevs_list": [ 00:17:59.224 { 00:17:59.224 "name": null, 00:17:59.224 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:59.224 "is_configured": false, 00:17:59.224 "data_offset": 0, 00:17:59.224 "data_size": 7936 00:17:59.224 }, 00:17:59.224 { 00:17:59.224 "name": "BaseBdev2", 00:17:59.224 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:59.224 "is_configured": true, 00:17:59.224 "data_offset": 256, 00:17:59.224 "data_size": 7936 00:17:59.224 } 00:17:59.224 ] 00:17:59.224 }' 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:59.224 "name": "raid_bdev1", 00:17:59.224 "uuid": "f560271d-453c-4d6d-b434-7de6020c4db4", 00:17:59.224 "strip_size_kb": 0, 00:17:59.224 "state": "online", 00:17:59.224 "raid_level": "raid1", 00:17:59.224 "superblock": true, 00:17:59.224 "num_base_bdevs": 2, 00:17:59.224 "num_base_bdevs_discovered": 1, 00:17:59.224 "num_base_bdevs_operational": 1, 00:17:59.224 "base_bdevs_list": [ 00:17:59.224 { 00:17:59.224 "name": null, 00:17:59.224 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:59.224 "is_configured": false, 00:17:59.224 "data_offset": 0, 00:17:59.224 "data_size": 7936 00:17:59.224 }, 00:17:59.224 { 00:17:59.224 "name": "BaseBdev2", 00:17:59.224 "uuid": "f0c9d136-98ee-5cf6-854a-86c3307bc886", 00:17:59.224 "is_configured": true, 00:17:59.224 "data_offset": 256, 00:17:59.224 "data_size": 7936 00:17:59.224 } 00:17:59.224 ] 00:17:59.224 }' 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:59.224 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:59.482 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:59.482 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@784 -- # killprocess 86491 00:17:59.482 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 86491 ']' 00:17:59.482 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 86491 00:17:59.482 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:17:59.482 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:59.482 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 86491 00:17:59.482 killing process with pid 86491 00:17:59.482 Received shutdown signal, test time was about 60.000000 seconds 00:17:59.482 00:17:59.482 Latency(us) 00:17:59.482 [2024-11-20T16:05:57.733Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:59.482 [2024-11-20T16:05:57.733Z] =================================================================================================================== 00:17:59.482 [2024-11-20T16:05:57.733Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:17:59.482 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:59.482 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:59.482 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 86491' 00:17:59.482 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@973 -- # kill 86491 00:17:59.482 16:05:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@978 -- # wait 86491 00:17:59.482 [2024-11-20 16:05:57.505663] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:59.482 [2024-11-20 16:05:57.505768] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:59.482 [2024-11-20 16:05:57.505808] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:59.482 [2024-11-20 16:05:57.505818] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:17:59.482 [2024-11-20 16:05:57.646352] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:18:00.048 16:05:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@786 -- # return 0 00:18:00.048 00:18:00.048 real 0m14.625s 00:18:00.048 user 0m18.588s 00:18:00.048 sys 0m1.060s 00:18:00.048 16:05:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:00.048 16:05:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:00.048 ************************************ 00:18:00.048 END TEST raid_rebuild_test_sb_md_interleaved 00:18:00.048 ************************************ 00:18:00.048 16:05:58 bdev_raid -- bdev/bdev_raid.sh@1015 -- # trap - EXIT 00:18:00.048 16:05:58 bdev_raid -- bdev/bdev_raid.sh@1016 -- # cleanup 00:18:00.048 16:05:58 bdev_raid -- bdev/bdev_raid.sh@56 -- # '[' -n 86491 ']' 00:18:00.048 16:05:58 bdev_raid -- bdev/bdev_raid.sh@56 -- # ps -p 86491 00:18:00.048 16:05:58 bdev_raid -- bdev/bdev_raid.sh@60 -- # rm -rf /raidtest 00:18:00.048 00:18:00.048 real 9m41.646s 00:18:00.048 user 12m50.423s 00:18:00.048 sys 1m20.371s 00:18:00.048 16:05:58 bdev_raid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:00.048 16:05:58 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:18:00.048 ************************************ 00:18:00.048 END TEST bdev_raid 00:18:00.048 ************************************ 00:18:00.048 16:05:58 -- spdk/autotest.sh@190 -- # run_test spdkcli_raid /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:18:00.048 16:05:58 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:18:00.048 16:05:58 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:00.048 16:05:58 -- common/autotest_common.sh@10 -- # set +x 00:18:00.048 ************************************ 00:18:00.048 START TEST spdkcli_raid 00:18:00.048 ************************************ 00:18:00.048 16:05:58 spdkcli_raid -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:18:00.307 * Looking for test storage... 00:18:00.307 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:18:00.307 16:05:58 spdkcli_raid -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:18:00.307 16:05:58 spdkcli_raid -- common/autotest_common.sh@1693 -- # lcov --version 00:18:00.307 16:05:58 spdkcli_raid -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:18:00.307 16:05:58 spdkcli_raid -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@336 -- # IFS=.-: 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@336 -- # read -ra ver1 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@337 -- # IFS=.-: 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@337 -- # read -ra ver2 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@338 -- # local 'op=<' 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@340 -- # ver1_l=2 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@341 -- # ver2_l=1 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@344 -- # case "$op" in 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@345 -- # : 1 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@365 -- # decimal 1 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@353 -- # local d=1 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@355 -- # echo 1 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@366 -- # decimal 2 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@353 -- # local d=2 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@355 -- # echo 2 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:18:00.307 16:05:58 spdkcli_raid -- scripts/common.sh@368 -- # return 0 00:18:00.307 16:05:58 spdkcli_raid -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:18:00.307 16:05:58 spdkcli_raid -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:18:00.307 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:00.307 --rc genhtml_branch_coverage=1 00:18:00.307 --rc genhtml_function_coverage=1 00:18:00.307 --rc genhtml_legend=1 00:18:00.307 --rc geninfo_all_blocks=1 00:18:00.307 --rc geninfo_unexecuted_blocks=1 00:18:00.307 00:18:00.307 ' 00:18:00.307 16:05:58 spdkcli_raid -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:18:00.307 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:00.307 --rc genhtml_branch_coverage=1 00:18:00.307 --rc genhtml_function_coverage=1 00:18:00.307 --rc genhtml_legend=1 00:18:00.307 --rc geninfo_all_blocks=1 00:18:00.307 --rc geninfo_unexecuted_blocks=1 00:18:00.307 00:18:00.307 ' 00:18:00.307 16:05:58 spdkcli_raid -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:18:00.307 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:00.307 --rc genhtml_branch_coverage=1 00:18:00.307 --rc genhtml_function_coverage=1 00:18:00.307 --rc genhtml_legend=1 00:18:00.307 --rc geninfo_all_blocks=1 00:18:00.307 --rc geninfo_unexecuted_blocks=1 00:18:00.307 00:18:00.307 ' 00:18:00.307 16:05:58 spdkcli_raid -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:18:00.307 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:00.307 --rc genhtml_branch_coverage=1 00:18:00.307 --rc genhtml_function_coverage=1 00:18:00.307 --rc genhtml_legend=1 00:18:00.307 --rc geninfo_all_blocks=1 00:18:00.307 --rc geninfo_unexecuted_blocks=1 00:18:00.307 00:18:00.307 ' 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/raid.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/raid.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/iscsi_tgt/common.sh 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@9 -- # ISCSI_BRIDGE=iscsi_br 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@10 -- # INITIATOR_INTERFACE=spdk_init_int 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@11 -- # INITIATOR_BRIDGE=init_br 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@12 -- # TARGET_NAMESPACE=spdk_iscsi_ns 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@13 -- # TARGET_NS_CMD=(ip netns exec "$TARGET_NAMESPACE") 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@14 -- # TARGET_INTERFACE=spdk_tgt_int 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@15 -- # TARGET_INTERFACE2=spdk_tgt_int2 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@16 -- # TARGET_BRIDGE=tgt_br 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@17 -- # TARGET_BRIDGE2=tgt_br2 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@20 -- # TARGET_IP=10.0.0.1 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@21 -- # TARGET_IP2=10.0.0.3 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@22 -- # INITIATOR_IP=10.0.0.2 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@23 -- # ISCSI_PORT=3260 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@24 -- # NETMASK=10.0.0.2/32 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@25 -- # INITIATOR_TAG=2 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@26 -- # INITIATOR_NAME=ANY 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@27 -- # PORTAL_TAG=1 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@28 -- # ISCSI_APP=("${TARGET_NS_CMD[@]}" "${ISCSI_APP[@]}") 00:18:00.307 16:05:58 spdkcli_raid -- iscsi_tgt/common.sh@29 -- # ISCSI_TEST_CORE_MASK=0xF 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/raid.sh@12 -- # MATCH_FILE=spdkcli_raid.test 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/raid.sh@13 -- # SPDKCLI_BRANCH=/bdevs 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/raid.sh@14 -- # dirname /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/raid.sh@14 -- # readlink -f /home/vagrant/spdk_repo/spdk/test/spdkcli 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/raid.sh@14 -- # testdir=/home/vagrant/spdk_repo/spdk/test/spdkcli 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/raid.sh@15 -- # . /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/raid.sh@17 -- # trap cleanup EXIT 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/raid.sh@19 -- # timing_enter run_spdk_tgt 00:18:00.307 16:05:58 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:18:00.307 16:05:58 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:00.307 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/raid.sh@20 -- # run_spdk_tgt 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/common.sh@27 -- # spdk_tgt_pid=87139 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/common.sh@28 -- # waitforlisten 87139 00:18:00.307 16:05:58 spdkcli_raid -- common/autotest_common.sh@835 -- # '[' -z 87139 ']' 00:18:00.307 16:05:58 spdkcli_raid -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:00.307 16:05:58 spdkcli_raid -- common/autotest_common.sh@840 -- # local max_retries=100 00:18:00.307 16:05:58 spdkcli_raid -- spdkcli/common.sh@26 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:18:00.307 16:05:58 spdkcli_raid -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:00.307 16:05:58 spdkcli_raid -- common/autotest_common.sh@844 -- # xtrace_disable 00:18:00.307 16:05:58 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:00.307 [2024-11-20 16:05:58.493406] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:18:00.308 [2024-11-20 16:05:58.493678] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87139 ] 00:18:00.566 [2024-11-20 16:05:58.651818] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:18:00.566 [2024-11-20 16:05:58.758568] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:18:00.566 [2024-11-20 16:05:58.758653] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:01.151 16:05:59 spdkcli_raid -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:18:01.151 16:05:59 spdkcli_raid -- common/autotest_common.sh@868 -- # return 0 00:18:01.151 16:05:59 spdkcli_raid -- spdkcli/raid.sh@21 -- # timing_exit run_spdk_tgt 00:18:01.151 16:05:59 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:01.151 16:05:59 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:01.151 16:05:59 spdkcli_raid -- spdkcli/raid.sh@23 -- # timing_enter spdkcli_create_malloc 00:18:01.151 16:05:59 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:18:01.151 16:05:59 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:01.151 16:05:59 spdkcli_raid -- spdkcli/raid.sh@26 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 8 512 Malloc1'\'' '\''Malloc1'\'' True 00:18:01.151 '\''/bdevs/malloc create 8 512 Malloc2'\'' '\''Malloc2'\'' True 00:18:01.151 ' 00:18:02.550 Executing command: ['/bdevs/malloc create 8 512 Malloc1', 'Malloc1', True] 00:18:02.550 Executing command: ['/bdevs/malloc create 8 512 Malloc2', 'Malloc2', True] 00:18:02.808 16:06:00 spdkcli_raid -- spdkcli/raid.sh@27 -- # timing_exit spdkcli_create_malloc 00:18:02.808 16:06:00 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:02.808 16:06:00 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:02.808 16:06:00 spdkcli_raid -- spdkcli/raid.sh@29 -- # timing_enter spdkcli_create_raid 00:18:02.808 16:06:00 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:18:02.808 16:06:00 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:02.808 16:06:00 spdkcli_raid -- spdkcli/raid.sh@31 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4'\'' '\''testraid'\'' True 00:18:02.808 ' 00:18:03.746 Executing command: ['/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4', 'testraid', True] 00:18:04.004 16:06:02 spdkcli_raid -- spdkcli/raid.sh@32 -- # timing_exit spdkcli_create_raid 00:18:04.004 16:06:02 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:04.004 16:06:02 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:04.004 16:06:02 spdkcli_raid -- spdkcli/raid.sh@34 -- # timing_enter spdkcli_check_match 00:18:04.004 16:06:02 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:18:04.004 16:06:02 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:04.004 16:06:02 spdkcli_raid -- spdkcli/raid.sh@35 -- # check_match 00:18:04.004 16:06:02 spdkcli_raid -- spdkcli/common.sh@44 -- # /home/vagrant/spdk_repo/spdk/scripts/spdkcli.py ll /bdevs 00:18:04.570 16:06:02 spdkcli_raid -- spdkcli/common.sh@45 -- # /home/vagrant/spdk_repo/spdk/test/app/match/match /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test.match 00:18:04.570 16:06:02 spdkcli_raid -- spdkcli/common.sh@46 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test 00:18:04.570 16:06:02 spdkcli_raid -- spdkcli/raid.sh@36 -- # timing_exit spdkcli_check_match 00:18:04.570 16:06:02 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:04.570 16:06:02 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:04.570 16:06:02 spdkcli_raid -- spdkcli/raid.sh@38 -- # timing_enter spdkcli_delete_raid 00:18:04.570 16:06:02 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:18:04.570 16:06:02 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:04.570 16:06:02 spdkcli_raid -- spdkcli/raid.sh@40 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume delete testraid'\'' '\'''\'' True 00:18:04.570 ' 00:18:05.504 Executing command: ['/bdevs/raid_volume delete testraid', '', True] 00:18:05.504 16:06:03 spdkcli_raid -- spdkcli/raid.sh@41 -- # timing_exit spdkcli_delete_raid 00:18:05.504 16:06:03 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:05.504 16:06:03 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:05.504 16:06:03 spdkcli_raid -- spdkcli/raid.sh@43 -- # timing_enter spdkcli_delete_malloc 00:18:05.504 16:06:03 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:18:05.504 16:06:03 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:05.504 16:06:03 spdkcli_raid -- spdkcli/raid.sh@46 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc delete Malloc1'\'' '\'''\'' True 00:18:05.504 '\''/bdevs/malloc delete Malloc2'\'' '\'''\'' True 00:18:05.504 ' 00:18:06.878 Executing command: ['/bdevs/malloc delete Malloc1', '', True] 00:18:06.878 Executing command: ['/bdevs/malloc delete Malloc2', '', True] 00:18:06.878 16:06:05 spdkcli_raid -- spdkcli/raid.sh@47 -- # timing_exit spdkcli_delete_malloc 00:18:06.878 16:06:05 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:06.878 16:06:05 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:06.878 16:06:05 spdkcli_raid -- spdkcli/raid.sh@49 -- # killprocess 87139 00:18:06.878 16:06:05 spdkcli_raid -- common/autotest_common.sh@954 -- # '[' -z 87139 ']' 00:18:06.878 16:06:05 spdkcli_raid -- common/autotest_common.sh@958 -- # kill -0 87139 00:18:06.878 16:06:05 spdkcli_raid -- common/autotest_common.sh@959 -- # uname 00:18:06.878 16:06:05 spdkcli_raid -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:18:06.878 16:06:05 spdkcli_raid -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87139 00:18:07.136 16:06:05 spdkcli_raid -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:18:07.136 16:06:05 spdkcli_raid -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:18:07.136 16:06:05 spdkcli_raid -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87139' 00:18:07.136 killing process with pid 87139 00:18:07.136 16:06:05 spdkcli_raid -- common/autotest_common.sh@973 -- # kill 87139 00:18:07.136 16:06:05 spdkcli_raid -- common/autotest_common.sh@978 -- # wait 87139 00:18:08.131 16:06:06 spdkcli_raid -- spdkcli/raid.sh@1 -- # cleanup 00:18:08.131 16:06:06 spdkcli_raid -- spdkcli/common.sh@10 -- # '[' -n 87139 ']' 00:18:08.131 16:06:06 spdkcli_raid -- spdkcli/common.sh@11 -- # killprocess 87139 00:18:08.131 16:06:06 spdkcli_raid -- common/autotest_common.sh@954 -- # '[' -z 87139 ']' 00:18:08.131 16:06:06 spdkcli_raid -- common/autotest_common.sh@958 -- # kill -0 87139 00:18:08.131 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (87139) - No such process 00:18:08.131 Process with pid 87139 is not found 00:18:08.131 16:06:06 spdkcli_raid -- common/autotest_common.sh@981 -- # echo 'Process with pid 87139 is not found' 00:18:08.131 16:06:06 spdkcli_raid -- spdkcli/common.sh@13 -- # '[' -n '' ']' 00:18:08.131 16:06:06 spdkcli_raid -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:18:08.131 16:06:06 spdkcli_raid -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:18:08.131 16:06:06 spdkcli_raid -- spdkcli/common.sh@22 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_raid.test /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:18:08.131 ************************************ 00:18:08.131 END TEST spdkcli_raid 00:18:08.131 ************************************ 00:18:08.131 00:18:08.131 real 0m8.049s 00:18:08.131 user 0m16.762s 00:18:08.131 sys 0m0.715s 00:18:08.131 16:06:06 spdkcli_raid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:08.131 16:06:06 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:08.131 16:06:06 -- spdk/autotest.sh@191 -- # run_test blockdev_raid5f /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:18:08.131 16:06:06 -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:18:08.131 16:06:06 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:08.131 16:06:06 -- common/autotest_common.sh@10 -- # set +x 00:18:08.131 ************************************ 00:18:08.131 START TEST blockdev_raid5f 00:18:08.131 ************************************ 00:18:08.131 16:06:06 blockdev_raid5f -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:18:08.390 * Looking for test storage... 00:18:08.390 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:18:08.390 16:06:06 blockdev_raid5f -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:18:08.390 16:06:06 blockdev_raid5f -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:18:08.390 16:06:06 blockdev_raid5f -- common/autotest_common.sh@1693 -- # lcov --version 00:18:08.390 16:06:06 blockdev_raid5f -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@333 -- # local ver1 ver1_l 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@334 -- # local ver2 ver2_l 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@336 -- # IFS=.-: 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@336 -- # read -ra ver1 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@337 -- # IFS=.-: 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@337 -- # read -ra ver2 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@338 -- # local 'op=<' 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@340 -- # ver1_l=2 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@341 -- # ver2_l=1 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@344 -- # case "$op" in 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@345 -- # : 1 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@364 -- # (( v = 0 )) 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@365 -- # decimal 1 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@353 -- # local d=1 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@355 -- # echo 1 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@365 -- # ver1[v]=1 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@366 -- # decimal 2 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@353 -- # local d=2 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@355 -- # echo 2 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@366 -- # ver2[v]=2 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:18:08.390 16:06:06 blockdev_raid5f -- scripts/common.sh@368 -- # return 0 00:18:08.390 16:06:06 blockdev_raid5f -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:18:08.390 16:06:06 blockdev_raid5f -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:18:08.390 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:08.390 --rc genhtml_branch_coverage=1 00:18:08.390 --rc genhtml_function_coverage=1 00:18:08.390 --rc genhtml_legend=1 00:18:08.390 --rc geninfo_all_blocks=1 00:18:08.390 --rc geninfo_unexecuted_blocks=1 00:18:08.390 00:18:08.390 ' 00:18:08.390 16:06:06 blockdev_raid5f -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:18:08.390 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:08.390 --rc genhtml_branch_coverage=1 00:18:08.390 --rc genhtml_function_coverage=1 00:18:08.390 --rc genhtml_legend=1 00:18:08.390 --rc geninfo_all_blocks=1 00:18:08.390 --rc geninfo_unexecuted_blocks=1 00:18:08.390 00:18:08.390 ' 00:18:08.390 16:06:06 blockdev_raid5f -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:18:08.390 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:08.390 --rc genhtml_branch_coverage=1 00:18:08.390 --rc genhtml_function_coverage=1 00:18:08.390 --rc genhtml_legend=1 00:18:08.390 --rc geninfo_all_blocks=1 00:18:08.390 --rc geninfo_unexecuted_blocks=1 00:18:08.390 00:18:08.390 ' 00:18:08.390 16:06:06 blockdev_raid5f -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:18:08.390 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:08.390 --rc genhtml_branch_coverage=1 00:18:08.390 --rc genhtml_function_coverage=1 00:18:08.390 --rc genhtml_legend=1 00:18:08.390 --rc geninfo_all_blocks=1 00:18:08.390 --rc geninfo_unexecuted_blocks=1 00:18:08.390 00:18:08.390 ' 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/nbd_common.sh@6 -- # set -e 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@12 -- # rpc_py=rpc_cmd 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@13 -- # conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@14 -- # nonenclosed_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@15 -- # nonarray_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@17 -- # export RPC_PIPE_TIMEOUT=30 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@17 -- # RPC_PIPE_TIMEOUT=30 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@20 -- # : 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@707 -- # QOS_DEV_1=Malloc_0 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@708 -- # QOS_DEV_2=Null_1 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@709 -- # QOS_RUN_TIME=5 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@711 -- # uname -s 00:18:08.390 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@711 -- # '[' Linux = Linux ']' 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@713 -- # PRE_RESERVED_MEM=0 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@719 -- # test_type=raid5f 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@720 -- # crypto_device= 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@721 -- # dek= 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@722 -- # env_ctx= 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@723 -- # wait_for_rpc= 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@724 -- # '[' -n '' ']' 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@727 -- # [[ raid5f == bdev ]] 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@727 -- # [[ raid5f == crypto_* ]] 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@730 -- # start_spdk_tgt 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@47 -- # spdk_tgt_pid=87397 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@48 -- # trap 'killprocess "$spdk_tgt_pid"; exit 1' SIGINT SIGTERM EXIT 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@49 -- # waitforlisten 87397 00:18:08.390 16:06:06 blockdev_raid5f -- common/autotest_common.sh@835 -- # '[' -z 87397 ']' 00:18:08.390 16:06:06 blockdev_raid5f -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:08.390 16:06:06 blockdev_raid5f -- common/autotest_common.sh@840 -- # local max_retries=100 00:18:08.390 16:06:06 blockdev_raid5f -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:08.390 16:06:06 blockdev_raid5f -- common/autotest_common.sh@844 -- # xtrace_disable 00:18:08.390 16:06:06 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:08.390 16:06:06 blockdev_raid5f -- bdev/blockdev.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt '' '' 00:18:08.390 [2024-11-20 16:06:06.588575] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:18:08.390 [2024-11-20 16:06:06.588685] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87397 ] 00:18:08.646 [2024-11-20 16:06:06.739665] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:08.646 [2024-11-20 16:06:06.820763] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:09.210 16:06:07 blockdev_raid5f -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:18:09.210 16:06:07 blockdev_raid5f -- common/autotest_common.sh@868 -- # return 0 00:18:09.210 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@731 -- # case "$test_type" in 00:18:09.210 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@763 -- # setup_raid5f_conf 00:18:09.210 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@279 -- # rpc_cmd 00:18:09.210 16:06:07 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:09.210 16:06:07 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:09.467 Malloc0 00:18:09.467 Malloc1 00:18:09.467 Malloc2 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:09.467 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@774 -- # rpc_cmd bdev_wait_for_examine 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:09.467 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@777 -- # cat 00:18:09.467 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n accel 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:09.467 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n bdev 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:09.467 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n iobuf 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:09.467 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@785 -- # mapfile -t bdevs 00:18:09.467 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@785 -- # jq -r '.[] | select(.claimed == false)' 00:18:09.467 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@785 -- # rpc_cmd bdev_get_bdevs 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:09.467 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@786 -- # mapfile -t bdevs_name 00:18:09.467 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@786 -- # jq -r .name 00:18:09.467 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@786 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "74c9fd4f-8952-4f1f-a560-7ba63e8b057d"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "74c9fd4f-8952-4f1f-a560-7ba63e8b057d",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "74c9fd4f-8952-4f1f-a560-7ba63e8b057d",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "74c218ba-8869-497e-bd3e-2ec574ab53d7",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "a0ba0921-b4c1-4197-b7a7-34d43815af41",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "05f733c9-4edf-4a75-a57e-b0d6ad280efc",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:18:09.467 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@787 -- # bdev_list=("${bdevs_name[@]}") 00:18:09.467 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@789 -- # hello_world_bdev=raid5f 00:18:09.467 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@790 -- # trap - SIGINT SIGTERM EXIT 00:18:09.467 16:06:07 blockdev_raid5f -- bdev/blockdev.sh@791 -- # killprocess 87397 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@954 -- # '[' -z 87397 ']' 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@958 -- # kill -0 87397 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@959 -- # uname 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87397 00:18:09.467 killing process with pid 87397 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:18:09.467 16:06:07 blockdev_raid5f -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:18:09.468 16:06:07 blockdev_raid5f -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87397' 00:18:09.468 16:06:07 blockdev_raid5f -- common/autotest_common.sh@973 -- # kill 87397 00:18:09.468 16:06:07 blockdev_raid5f -- common/autotest_common.sh@978 -- # wait 87397 00:18:10.840 16:06:08 blockdev_raid5f -- bdev/blockdev.sh@795 -- # trap cleanup SIGINT SIGTERM EXIT 00:18:10.840 16:06:08 blockdev_raid5f -- bdev/blockdev.sh@797 -- # run_test bdev_hello_world /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:18:10.840 16:06:08 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:18:10.840 16:06:08 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:10.840 16:06:08 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:10.840 ************************************ 00:18:10.840 START TEST bdev_hello_world 00:18:10.840 ************************************ 00:18:10.840 16:06:08 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:18:10.840 [2024-11-20 16:06:09.041468] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:18:10.840 [2024-11-20 16:06:09.041708] [ DPDK EAL parameters: hello_bdev --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87448 ] 00:18:11.100 [2024-11-20 16:06:09.191032] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:11.100 [2024-11-20 16:06:09.271428] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:11.359 [2024-11-20 16:06:09.605550] hello_bdev.c: 222:hello_start: *NOTICE*: Successfully started the application 00:18:11.359 [2024-11-20 16:06:09.605592] hello_bdev.c: 231:hello_start: *NOTICE*: Opening the bdev raid5f 00:18:11.359 [2024-11-20 16:06:09.605605] hello_bdev.c: 244:hello_start: *NOTICE*: Opening io channel 00:18:11.359 [2024-11-20 16:06:09.605979] hello_bdev.c: 138:hello_write: *NOTICE*: Writing to the bdev 00:18:11.359 [2024-11-20 16:06:09.606067] hello_bdev.c: 117:write_complete: *NOTICE*: bdev io write completed successfully 00:18:11.359 [2024-11-20 16:06:09.606082] hello_bdev.c: 84:hello_read: *NOTICE*: Reading io 00:18:11.359 [2024-11-20 16:06:09.606123] hello_bdev.c: 65:read_complete: *NOTICE*: Read string from bdev : Hello World! 00:18:11.359 00:18:11.359 [2024-11-20 16:06:09.606136] hello_bdev.c: 74:read_complete: *NOTICE*: Stopping app 00:18:12.293 00:18:12.293 real 0m1.310s 00:18:12.293 user 0m1.030s 00:18:12.293 sys 0m0.163s 00:18:12.293 16:06:10 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:12.293 16:06:10 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@10 -- # set +x 00:18:12.293 ************************************ 00:18:12.293 END TEST bdev_hello_world 00:18:12.293 ************************************ 00:18:12.293 16:06:10 blockdev_raid5f -- bdev/blockdev.sh@798 -- # run_test bdev_bounds bdev_bounds '' 00:18:12.293 16:06:10 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:18:12.293 16:06:10 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:12.293 16:06:10 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:12.293 ************************************ 00:18:12.293 START TEST bdev_bounds 00:18:12.293 ************************************ 00:18:12.293 16:06:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1129 -- # bdev_bounds '' 00:18:12.293 16:06:10 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@289 -- # bdevio_pid=87480 00:18:12.293 16:06:10 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@290 -- # trap 'cleanup; killprocess $bdevio_pid; exit 1' SIGINT SIGTERM EXIT 00:18:12.293 16:06:10 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@288 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/bdevio -w -s 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:18:12.293 Process bdevio pid: 87480 00:18:12.293 16:06:10 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@291 -- # echo 'Process bdevio pid: 87480' 00:18:12.293 16:06:10 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@292 -- # waitforlisten 87480 00:18:12.293 16:06:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@835 -- # '[' -z 87480 ']' 00:18:12.293 16:06:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:12.293 16:06:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@840 -- # local max_retries=100 00:18:12.293 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:12.293 16:06:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:12.293 16:06:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@844 -- # xtrace_disable 00:18:12.293 16:06:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:18:12.293 [2024-11-20 16:06:10.397408] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:18:12.293 [2024-11-20 16:06:10.397509] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 0 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87480 ] 00:18:12.551 [2024-11-20 16:06:10.546795] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:18:12.551 [2024-11-20 16:06:10.629562] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:18:12.551 [2024-11-20 16:06:10.629835] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:18:12.551 [2024-11-20 16:06:10.629850] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:13.117 16:06:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:18:13.117 16:06:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@868 -- # return 0 00:18:13.117 16:06:11 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@293 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/tests.py perform_tests 00:18:13.117 I/O targets: 00:18:13.117 raid5f: 131072 blocks of 512 bytes (64 MiB) 00:18:13.117 00:18:13.117 00:18:13.117 CUnit - A unit testing framework for C - Version 2.1-3 00:18:13.117 http://cunit.sourceforge.net/ 00:18:13.117 00:18:13.117 00:18:13.117 Suite: bdevio tests on: raid5f 00:18:13.117 Test: blockdev write read block ...passed 00:18:13.117 Test: blockdev write zeroes read block ...passed 00:18:13.117 Test: blockdev write zeroes read no split ...passed 00:18:13.374 Test: blockdev write zeroes read split ...passed 00:18:13.374 Test: blockdev write zeroes read split partial ...passed 00:18:13.374 Test: blockdev reset ...passed 00:18:13.374 Test: blockdev write read 8 blocks ...passed 00:18:13.374 Test: blockdev write read size > 128k ...passed 00:18:13.374 Test: blockdev write read invalid size ...passed 00:18:13.374 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:18:13.374 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:18:13.374 Test: blockdev write read max offset ...passed 00:18:13.374 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:18:13.374 Test: blockdev writev readv 8 blocks ...passed 00:18:13.374 Test: blockdev writev readv 30 x 1block ...passed 00:18:13.374 Test: blockdev writev readv block ...passed 00:18:13.374 Test: blockdev writev readv size > 128k ...passed 00:18:13.374 Test: blockdev writev readv size > 128k in two iovs ...passed 00:18:13.374 Test: blockdev comparev and writev ...passed 00:18:13.374 Test: blockdev nvme passthru rw ...passed 00:18:13.374 Test: blockdev nvme passthru vendor specific ...passed 00:18:13.374 Test: blockdev nvme admin passthru ...passed 00:18:13.374 Test: blockdev copy ...passed 00:18:13.374 00:18:13.374 Run Summary: Type Total Ran Passed Failed Inactive 00:18:13.375 suites 1 1 n/a 0 0 00:18:13.375 tests 23 23 23 0 0 00:18:13.375 asserts 130 130 130 0 n/a 00:18:13.375 00:18:13.375 Elapsed time = 0.437 seconds 00:18:13.375 0 00:18:13.375 16:06:11 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@294 -- # killprocess 87480 00:18:13.375 16:06:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@954 -- # '[' -z 87480 ']' 00:18:13.375 16:06:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@958 -- # kill -0 87480 00:18:13.375 16:06:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@959 -- # uname 00:18:13.375 16:06:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:18:13.375 16:06:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87480 00:18:13.375 16:06:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:18:13.375 killing process with pid 87480 00:18:13.375 16:06:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:18:13.375 16:06:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87480' 00:18:13.375 16:06:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@973 -- # kill 87480 00:18:13.375 16:06:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@978 -- # wait 87480 00:18:14.309 16:06:12 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@295 -- # trap - SIGINT SIGTERM EXIT 00:18:14.309 00:18:14.309 real 0m1.943s 00:18:14.309 user 0m4.900s 00:18:14.309 sys 0m0.284s 00:18:14.309 16:06:12 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:14.310 ************************************ 00:18:14.310 END TEST bdev_bounds 00:18:14.310 ************************************ 00:18:14.310 16:06:12 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:18:14.310 16:06:12 blockdev_raid5f -- bdev/blockdev.sh@799 -- # run_test bdev_nbd nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:18:14.310 16:06:12 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:18:14.310 16:06:12 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:14.310 16:06:12 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:14.310 ************************************ 00:18:14.310 START TEST bdev_nbd 00:18:14.310 ************************************ 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1129 -- # nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # uname -s 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # [[ Linux == Linux ]] 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@301 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@302 -- # local conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # bdev_all=('raid5f') 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # local bdev_all 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@304 -- # local bdev_num=1 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@308 -- # [[ -e /sys/module/nbd ]] 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # nbd_all=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # local nbd_all 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@311 -- # bdev_num=1 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # nbd_list=('/dev/nbd0') 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # local nbd_list 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # bdev_list=('raid5f') 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # local bdev_list 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@317 -- # nbd_pid=87533 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@318 -- # trap 'cleanup; killprocess $nbd_pid' SIGINT SIGTERM EXIT 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@319 -- # waitforlisten 87533 /var/tmp/spdk-nbd.sock 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@835 -- # '[' -z 87533 ']' 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@840 -- # local max_retries=100 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:18:14.310 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@844 -- # xtrace_disable 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:18:14.310 16:06:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@316 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-nbd.sock -i 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:18:14.310 [2024-11-20 16:06:12.387552] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:18:14.310 [2024-11-20 16:06:12.387641] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:18:14.310 [2024-11-20 16:06:12.534369] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:14.568 [2024-11-20 16:06:12.616134] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:15.135 16:06:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:18:15.135 16:06:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@868 -- # return 0 00:18:15.135 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@321 -- # nbd_rpc_start_stop_verify /var/tmp/spdk-nbd.sock raid5f 00:18:15.135 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@113 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:15.135 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # bdev_list=('raid5f') 00:18:15.135 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # local bdev_list 00:18:15.135 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@116 -- # nbd_start_disks_without_nbd_idx /var/tmp/spdk-nbd.sock raid5f 00:18:15.135 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@22 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:15.135 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # bdev_list=('raid5f') 00:18:15.135 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # local bdev_list 00:18:15.135 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@24 -- # local i 00:18:15.135 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@25 -- # local nbd_device 00:18:15.135 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i = 0 )) 00:18:15.135 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:18:15.135 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd0 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd0 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd0 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # local i 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@877 -- # break 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:18:15.394 1+0 records in 00:18:15.394 1+0 records out 00:18:15.394 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000224242 s, 18.3 MB/s 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # size=4096 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@893 -- # return 0 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # nbd_disks_json='[ 00:18:15.394 { 00:18:15.394 "nbd_device": "/dev/nbd0", 00:18:15.394 "bdev_name": "raid5f" 00:18:15.394 } 00:18:15.394 ]' 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # nbd_disks_name=($(echo "${nbd_disks_json}" | jq -r '.[] | .nbd_device')) 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # jq -r '.[] | .nbd_device' 00:18:15.394 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # echo '[ 00:18:15.394 { 00:18:15.394 "nbd_device": "/dev/nbd0", 00:18:15.394 "bdev_name": "raid5f" 00:18:15.394 } 00:18:15.394 ]' 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@120 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:15.653 16:06:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # count=0 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@123 -- # '[' 0 -ne 0 ']' 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@127 -- # return 0 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@322 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # bdev_list=('raid5f') 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # local bdev_list 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0') 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # local nbd_list 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # bdev_list=('raid5f') 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # local bdev_list 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # local nbd_list 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@12 -- # local i 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:18:15.911 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f /dev/nbd0 00:18:16.169 /dev/nbd0 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # local i 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@877 -- # break 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:18:16.169 1+0 records in 00:18:16.169 1+0 records out 00:18:16.169 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000349535 s, 11.7 MB/s 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # size=4096 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@893 -- # return 0 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:16.169 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:18:16.430 { 00:18:16.430 "nbd_device": "/dev/nbd0", 00:18:16.430 "bdev_name": "raid5f" 00:18:16.430 } 00:18:16.430 ]' 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[ 00:18:16.430 { 00:18:16.430 "nbd_device": "/dev/nbd0", 00:18:16.430 "bdev_name": "raid5f" 00:18:16.430 } 00:18:16.430 ]' 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=1 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 1 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # count=1 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@96 -- # '[' 1 -ne 1 ']' 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify /dev/nbd0 write 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=write 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest bs=4096 count=256 00:18:16.430 256+0 records in 00:18:16.430 256+0 records out 00:18:16.430 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0135882 s, 77.2 MB/s 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:18:16.430 256+0 records in 00:18:16.430 256+0 records out 00:18:16.430 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0227136 s, 46.2 MB/s 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify /dev/nbd0 verify 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=verify 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest /dev/nbd0 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:18:16.430 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:18:16.689 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:18:16.690 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:18:16.690 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:18:16.690 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:18:16.690 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:18:16.690 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:18:16.690 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:18:16.690 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:18:16.690 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:18:16.690 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:16.690 16:06:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:18:16.948 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:18:16.948 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:18:16.948 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:18:16.948 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:18:16.948 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:18:16.948 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:18:16.948 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:18:16.948 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:18:16.948 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:18:16.948 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # count=0 00:18:16.948 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:18:16.948 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@109 -- # return 0 00:18:16.948 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@323 -- # nbd_with_lvol_verify /var/tmp/spdk-nbd.sock /dev/nbd0 00:18:16.948 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@131 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:16.948 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@132 -- # local nbd=/dev/nbd0 00:18:16.948 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@134 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create -b malloc_lvol_verify 16 512 00:18:17.207 malloc_lvol_verify 00:18:17.207 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@135 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create_lvstore malloc_lvol_verify lvs 00:18:17.207 c0fa972a-5e6e-4a3a-8287-21cb6c0d4672 00:18:17.465 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@136 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create lvol 4 -l lvs 00:18:17.465 0fa00bef-a88f-4a36-b093-d8c6b7168fd1 00:18:17.465 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@137 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk lvs/lvol /dev/nbd0 00:18:17.722 /dev/nbd0 00:18:17.722 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@139 -- # wait_for_nbd_set_capacity /dev/nbd0 00:18:17.722 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@146 -- # local nbd=nbd0 00:18:17.722 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@148 -- # [[ -e /sys/block/nbd0/size ]] 00:18:17.722 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@150 -- # (( 8192 == 0 )) 00:18:17.722 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@141 -- # mkfs.ext4 /dev/nbd0 00:18:17.722 mke2fs 1.47.0 (5-Feb-2023) 00:18:17.722 Discarding device blocks: 0/4096 done 00:18:17.722 Creating filesystem with 4096 1k blocks and 1024 inodes 00:18:17.722 00:18:17.722 Allocating group tables: 0/1 done 00:18:17.722 Writing inode tables: 0/1 done 00:18:17.722 Creating journal (1024 blocks): done 00:18:17.722 Writing superblocks and filesystem accounting information: 0/1 done 00:18:17.722 00:18:17.722 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@142 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:18:17.722 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:17.722 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:18:17.722 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:18:17.722 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:18:17.722 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:18:17.722 16:06:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@325 -- # killprocess 87533 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@954 -- # '[' -z 87533 ']' 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@958 -- # kill -0 87533 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@959 -- # uname 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87533 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:18:17.980 killing process with pid 87533 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87533' 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@973 -- # kill 87533 00:18:17.980 16:06:16 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@978 -- # wait 87533 00:18:18.913 16:06:17 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@326 -- # trap - SIGINT SIGTERM EXIT 00:18:18.913 00:18:18.913 real 0m4.720s 00:18:18.913 user 0m6.747s 00:18:18.913 sys 0m0.925s 00:18:18.913 16:06:17 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:18.914 16:06:17 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:18:18.914 ************************************ 00:18:18.914 END TEST bdev_nbd 00:18:18.914 ************************************ 00:18:18.914 16:06:17 blockdev_raid5f -- bdev/blockdev.sh@800 -- # [[ y == y ]] 00:18:18.914 16:06:17 blockdev_raid5f -- bdev/blockdev.sh@801 -- # '[' raid5f = nvme ']' 00:18:18.914 16:06:17 blockdev_raid5f -- bdev/blockdev.sh@801 -- # '[' raid5f = gpt ']' 00:18:18.914 16:06:17 blockdev_raid5f -- bdev/blockdev.sh@805 -- # run_test bdev_fio fio_test_suite '' 00:18:18.914 16:06:17 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:18:18.914 16:06:17 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:18.914 16:06:17 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:18.914 ************************************ 00:18:18.914 START TEST bdev_fio 00:18:18.914 ************************************ 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1129 -- # fio_test_suite '' 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@330 -- # local env_context 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@334 -- # pushd /home/vagrant/spdk_repo/spdk/test/bdev 00:18:18.914 /home/vagrant/spdk_repo/spdk/test/bdev /home/vagrant/spdk_repo/spdk 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@335 -- # trap 'rm -f ./*.state; popd; exit 1' SIGINT SIGTERM EXIT 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # echo '' 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # sed s/--env-context=// 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # env_context= 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@339 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio verify AIO '' 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1285 -- # local workload=verify 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # local bdev_type=AIO 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1287 -- # local env_context= 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1288 -- # local fio_dir=/usr/src/fio 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1290 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -z verify ']' 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # '[' -n '' ']' 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1303 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1305 -- # cat 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1317 -- # '[' verify == verify ']' 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1318 -- # cat 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1327 -- # '[' AIO == AIO ']' 00:18:18.914 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # /usr/src/fio/fio --version 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # [[ fio-3.35 == *\f\i\o\-\3* ]] 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1329 -- # echo serialize_overlap=1 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@340 -- # for b in "${bdevs_name[@]}" 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@341 -- # echo '[job_raid5f]' 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@342 -- # echo filename=raid5f 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@346 -- # local 'fio_params=--ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json' 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@348 -- # run_test bdev_fio_rw_verify fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1105 -- # '[' 11 -le 1 ']' 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:18:19.172 ************************************ 00:18:19.172 START TEST bdev_fio_rw_verify 00:18:19.172 ************************************ 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1129 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1360 -- # fio_plugin /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # local sanitizers 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # local plugin=/home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # shift 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1347 -- # local asan_lib= 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # ldd /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # grep libasan 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # asan_lib=/usr/lib64/libasan.so.8 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1350 -- # [[ -n /usr/lib64/libasan.so.8 ]] 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1351 -- # break 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # LD_PRELOAD='/usr/lib64/libasan.so.8 /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev' 00:18:19.172 16:06:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:18:19.172 job_raid5f: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:18:19.172 fio-3.35 00:18:19.172 Starting 1 thread 00:18:31.372 00:18:31.372 job_raid5f: (groupid=0, jobs=1): err= 0: pid=87721: Wed Nov 20 16:06:28 2024 00:18:31.372 read: IOPS=12.6k, BW=49.4MiB/s (51.8MB/s)(494MiB/10001msec) 00:18:31.372 slat (nsec): min=17522, max=44996, avg=19306.05, stdev=1981.37 00:18:31.372 clat (usec): min=8, max=314, avg=129.37, stdev=46.65 00:18:31.372 lat (usec): min=26, max=338, avg=148.68, stdev=47.07 00:18:31.372 clat percentiles (usec): 00:18:31.372 | 50.000th=[ 133], 99.000th=[ 245], 99.900th=[ 255], 99.990th=[ 277], 00:18:31.372 | 99.999th=[ 310] 00:18:31.372 write: IOPS=13.2k, BW=51.6MiB/s (54.1MB/s)(509MiB/9879msec); 0 zone resets 00:18:31.372 slat (usec): min=7, max=208, avg=16.04, stdev= 2.36 00:18:31.372 clat (usec): min=52, max=658, avg=288.82, stdev=40.83 00:18:31.372 lat (usec): min=67, max=759, avg=304.86, stdev=41.75 00:18:31.372 clat percentiles (usec): 00:18:31.372 | 50.000th=[ 293], 99.000th=[ 408], 99.900th=[ 429], 99.990th=[ 578], 00:18:31.372 | 99.999th=[ 627] 00:18:31.372 bw ( KiB/s): min=42104, max=56544, per=98.93%, avg=52239.16, stdev=3964.27, samples=19 00:18:31.372 iops : min=10526, max=14136, avg=13059.89, stdev=991.08, samples=19 00:18:31.372 lat (usec) : 10=0.01%, 20=0.01%, 50=0.01%, 100=17.07%, 250=40.83% 00:18:31.372 lat (usec) : 500=42.09%, 750=0.01% 00:18:31.372 cpu : usr=99.30%, sys=0.19%, ctx=28, majf=0, minf=10293 00:18:31.372 IO depths : 1=7.6%, 2=19.9%, 4=55.2%, 8=17.4%, 16=0.0%, 32=0.0%, >=64=0.0% 00:18:31.372 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:18:31.372 complete : 0=0.0%, 4=90.0%, 8=10.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:18:31.372 issued rwts: total=126361,130414,0,0 short=0,0,0,0 dropped=0,0,0,0 00:18:31.372 latency : target=0, window=0, percentile=100.00%, depth=8 00:18:31.372 00:18:31.372 Run status group 0 (all jobs): 00:18:31.372 READ: bw=49.4MiB/s (51.8MB/s), 49.4MiB/s-49.4MiB/s (51.8MB/s-51.8MB/s), io=494MiB (518MB), run=10001-10001msec 00:18:31.372 WRITE: bw=51.6MiB/s (54.1MB/s), 51.6MiB/s-51.6MiB/s (54.1MB/s-54.1MB/s), io=509MiB (534MB), run=9879-9879msec 00:18:31.372 ----------------------------------------------------- 00:18:31.372 Suppressions used: 00:18:31.372 count bytes template 00:18:31.372 1 7 /usr/src/fio/parse.c 00:18:31.372 133 12768 /usr/src/fio/iolog.c 00:18:31.372 1 8 libtcmalloc_minimal.so 00:18:31.372 1 904 libcrypto.so 00:18:31.372 ----------------------------------------------------- 00:18:31.373 00:18:31.373 00:18:31.373 real 0m12.061s 00:18:31.373 user 0m12.742s 00:18:31.373 sys 0m0.429s 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@10 -- # set +x 00:18:31.373 ************************************ 00:18:31.373 END TEST bdev_fio_rw_verify 00:18:31.373 ************************************ 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@349 -- # rm -f 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@350 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@353 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio trim '' '' 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1285 -- # local workload=trim 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # local bdev_type= 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1287 -- # local env_context= 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1288 -- # local fio_dir=/usr/src/fio 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1290 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -z trim ']' 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # '[' -n '' ']' 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1303 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1305 -- # cat 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1317 -- # '[' trim == verify ']' 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1332 -- # '[' trim == trim ']' 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1333 -- # echo rw=trimwrite 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "74c9fd4f-8952-4f1f-a560-7ba63e8b057d"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "74c9fd4f-8952-4f1f-a560-7ba63e8b057d",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "74c9fd4f-8952-4f1f-a560-7ba63e8b057d",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "74c218ba-8869-497e-bd3e-2ec574ab53d7",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "a0ba0921-b4c1-4197-b7a7-34d43815af41",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "05f733c9-4edf-4a75-a57e-b0d6ad280efc",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # [[ -n '' ]] 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@360 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@361 -- # popd 00:18:31.373 /home/vagrant/spdk_repo/spdk 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@362 -- # trap - SIGINT SIGTERM EXIT 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@363 -- # return 0 00:18:31.373 00:18:31.373 real 0m12.233s 00:18:31.373 user 0m12.820s 00:18:31.373 sys 0m0.499s 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:31.373 16:06:29 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:18:31.373 ************************************ 00:18:31.373 END TEST bdev_fio 00:18:31.373 ************************************ 00:18:31.373 16:06:29 blockdev_raid5f -- bdev/blockdev.sh@812 -- # trap cleanup SIGINT SIGTERM EXIT 00:18:31.373 16:06:29 blockdev_raid5f -- bdev/blockdev.sh@814 -- # run_test bdev_verify /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:18:31.373 16:06:29 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 16 -le 1 ']' 00:18:31.373 16:06:29 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:31.373 16:06:29 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:31.373 ************************************ 00:18:31.373 START TEST bdev_verify 00:18:31.373 ************************************ 00:18:31.373 16:06:29 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:18:31.373 [2024-11-20 16:06:29.416036] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:18:31.373 [2024-11-20 16:06:29.416163] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87890 ] 00:18:31.373 [2024-11-20 16:06:29.575527] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:18:31.631 [2024-11-20 16:06:29.678013] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:31.631 [2024-11-20 16:06:29.678026] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:18:31.890 Running I/O for 5 seconds... 00:18:34.197 16139.00 IOPS, 63.04 MiB/s [2024-11-20T16:06:33.383Z] 16601.00 IOPS, 64.85 MiB/s [2024-11-20T16:06:34.316Z] 17586.33 IOPS, 68.70 MiB/s [2024-11-20T16:06:35.252Z] 18192.00 IOPS, 71.06 MiB/s [2024-11-20T16:06:35.252Z] 18763.20 IOPS, 73.29 MiB/s 00:18:37.001 Latency(us) 00:18:37.001 [2024-11-20T16:06:35.252Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:37.001 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:18:37.001 Verification LBA range: start 0x0 length 0x2000 00:18:37.001 raid5f : 5.01 9298.87 36.32 0.00 0.00 20547.96 168.57 20568.22 00:18:37.001 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:18:37.001 Verification LBA range: start 0x2000 length 0x2000 00:18:37.001 raid5f : 5.01 9427.45 36.83 0.00 0.00 20248.50 163.84 20366.57 00:18:37.001 [2024-11-20T16:06:35.252Z] =================================================================================================================== 00:18:37.001 [2024-11-20T16:06:35.252Z] Total : 18726.32 73.15 0.00 0.00 20397.15 163.84 20568.22 00:18:37.934 00:18:37.934 real 0m6.471s 00:18:37.934 user 0m12.095s 00:18:37.934 sys 0m0.194s 00:18:37.934 16:06:35 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:37.934 ************************************ 00:18:37.934 END TEST bdev_verify 00:18:37.934 ************************************ 00:18:37.934 16:06:35 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@10 -- # set +x 00:18:37.934 16:06:35 blockdev_raid5f -- bdev/blockdev.sh@815 -- # run_test bdev_verify_big_io /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:18:37.934 16:06:35 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 16 -le 1 ']' 00:18:37.934 16:06:35 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:37.934 16:06:35 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:37.934 ************************************ 00:18:37.934 START TEST bdev_verify_big_io 00:18:37.934 ************************************ 00:18:37.934 16:06:35 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:18:37.934 [2024-11-20 16:06:35.921450] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:18:37.934 [2024-11-20 16:06:35.921548] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87977 ] 00:18:37.934 [2024-11-20 16:06:36.071857] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:18:37.934 [2024-11-20 16:06:36.156945] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:18:37.934 [2024-11-20 16:06:36.157162] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:38.500 Running I/O for 5 seconds... 00:18:40.377 1012.00 IOPS, 63.25 MiB/s [2024-11-20T16:06:40.000Z] 1046.00 IOPS, 65.38 MiB/s [2024-11-20T16:06:40.932Z] 1077.67 IOPS, 67.35 MiB/s [2024-11-20T16:06:41.866Z] 1062.25 IOPS, 66.39 MiB/s [2024-11-20T16:06:41.866Z] 1066.40 IOPS, 66.65 MiB/s 00:18:43.615 Latency(us) 00:18:43.615 [2024-11-20T16:06:41.866Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:43.615 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:18:43.615 Verification LBA range: start 0x0 length 0x200 00:18:43.615 raid5f : 5.16 565.78 35.36 0.00 0.00 5557293.28 133.12 274242.95 00:18:43.615 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:18:43.615 Verification LBA range: start 0x200 length 0x200 00:18:43.615 raid5f : 5.14 518.71 32.42 0.00 0.00 5983632.21 148.09 298440.86 00:18:43.615 [2024-11-20T16:06:41.866Z] =================================================================================================================== 00:18:43.615 [2024-11-20T16:06:41.866Z] Total : 1084.49 67.78 0.00 0.00 5760773.22 133.12 298440.86 00:18:44.181 00:18:44.181 real 0m6.546s 00:18:44.181 user 0m12.300s 00:18:44.181 sys 0m0.174s 00:18:44.181 16:06:42 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:44.181 16:06:42 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@10 -- # set +x 00:18:44.181 ************************************ 00:18:44.181 END TEST bdev_verify_big_io 00:18:44.181 ************************************ 00:18:44.439 16:06:42 blockdev_raid5f -- bdev/blockdev.sh@816 -- # run_test bdev_write_zeroes /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:18:44.439 16:06:42 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:18:44.439 16:06:42 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:44.439 16:06:42 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:44.439 ************************************ 00:18:44.439 START TEST bdev_write_zeroes 00:18:44.439 ************************************ 00:18:44.439 16:06:42 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:18:44.439 [2024-11-20 16:06:42.529656] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:18:44.439 [2024-11-20 16:06:42.529833] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88063 ] 00:18:44.697 [2024-11-20 16:06:42.697319] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:44.697 [2024-11-20 16:06:42.797040] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:44.983 Running I/O for 1 seconds... 00:18:46.357 22719.00 IOPS, 88.75 MiB/s 00:18:46.357 Latency(us) 00:18:46.357 [2024-11-20T16:06:44.609Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:46.358 Job: raid5f (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:18:46.358 raid5f : 1.01 22683.72 88.61 0.00 0.00 5623.01 1512.37 7612.26 00:18:46.358 [2024-11-20T16:06:44.609Z] =================================================================================================================== 00:18:46.358 [2024-11-20T16:06:44.609Z] Total : 22683.72 88.61 0.00 0.00 5623.01 1512.37 7612.26 00:18:46.922 00:18:46.922 real 0m2.652s 00:18:46.922 user 0m2.329s 00:18:46.922 sys 0m0.197s 00:18:46.922 16:06:45 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:46.922 ************************************ 00:18:46.922 END TEST bdev_write_zeroes 00:18:46.922 ************************************ 00:18:46.922 16:06:45 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@10 -- # set +x 00:18:46.922 16:06:45 blockdev_raid5f -- bdev/blockdev.sh@819 -- # run_test bdev_json_nonenclosed /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:18:46.922 16:06:45 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:18:46.922 16:06:45 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:46.922 16:06:45 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:46.922 ************************************ 00:18:46.922 START TEST bdev_json_nonenclosed 00:18:46.922 ************************************ 00:18:46.922 16:06:45 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:18:47.180 [2024-11-20 16:06:45.199848] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:18:47.180 [2024-11-20 16:06:45.199964] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88112 ] 00:18:47.180 [2024-11-20 16:06:45.357343] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:47.437 [2024-11-20 16:06:45.459426] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:47.437 [2024-11-20 16:06:45.459509] json_config.c: 608:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: not enclosed in {}. 00:18:47.437 [2024-11-20 16:06:45.459531] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:18:47.437 [2024-11-20 16:06:45.459541] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:18:47.437 00:18:47.437 real 0m0.499s 00:18:47.437 user 0m0.303s 00:18:47.437 sys 0m0.092s 00:18:47.437 16:06:45 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:47.437 16:06:45 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@10 -- # set +x 00:18:47.437 ************************************ 00:18:47.437 END TEST bdev_json_nonenclosed 00:18:47.437 ************************************ 00:18:47.437 16:06:45 blockdev_raid5f -- bdev/blockdev.sh@822 -- # run_test bdev_json_nonarray /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:18:47.437 16:06:45 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:18:47.437 16:06:45 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:47.437 16:06:45 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:47.437 ************************************ 00:18:47.437 START TEST bdev_json_nonarray 00:18:47.437 ************************************ 00:18:47.437 16:06:45 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:18:47.695 [2024-11-20 16:06:45.744369] Starting SPDK v25.01-pre git sha1 0728de5b0 / DPDK 24.03.0 initialization... 00:18:47.695 [2024-11-20 16:06:45.744476] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88138 ] 00:18:47.695 [2024-11-20 16:06:45.904470] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:47.953 [2024-11-20 16:06:46.004251] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:47.953 [2024-11-20 16:06:46.004337] json_config.c: 614:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: 'subsystems' should be an array. 00:18:47.953 [2024-11-20 16:06:46.004354] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:18:47.953 [2024-11-20 16:06:46.004369] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:18:47.953 00:18:47.953 real 0m0.504s 00:18:47.953 user 0m0.308s 00:18:47.953 sys 0m0.093s 00:18:47.953 16:06:46 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:47.953 16:06:46 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@10 -- # set +x 00:18:47.953 ************************************ 00:18:47.953 END TEST bdev_json_nonarray 00:18:47.953 ************************************ 00:18:48.212 16:06:46 blockdev_raid5f -- bdev/blockdev.sh@824 -- # [[ raid5f == bdev ]] 00:18:48.212 16:06:46 blockdev_raid5f -- bdev/blockdev.sh@832 -- # [[ raid5f == gpt ]] 00:18:48.212 16:06:46 blockdev_raid5f -- bdev/blockdev.sh@836 -- # [[ raid5f == crypto_sw ]] 00:18:48.212 16:06:46 blockdev_raid5f -- bdev/blockdev.sh@848 -- # trap - SIGINT SIGTERM EXIT 00:18:48.212 16:06:46 blockdev_raid5f -- bdev/blockdev.sh@849 -- # cleanup 00:18:48.212 16:06:46 blockdev_raid5f -- bdev/blockdev.sh@23 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/aiofile 00:18:48.212 16:06:46 blockdev_raid5f -- bdev/blockdev.sh@24 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:18:48.212 16:06:46 blockdev_raid5f -- bdev/blockdev.sh@26 -- # [[ raid5f == rbd ]] 00:18:48.212 16:06:46 blockdev_raid5f -- bdev/blockdev.sh@30 -- # [[ raid5f == daos ]] 00:18:48.212 16:06:46 blockdev_raid5f -- bdev/blockdev.sh@34 -- # [[ raid5f = \g\p\t ]] 00:18:48.212 16:06:46 blockdev_raid5f -- bdev/blockdev.sh@40 -- # [[ raid5f == xnvme ]] 00:18:48.212 ************************************ 00:18:48.212 END TEST blockdev_raid5f 00:18:48.212 ************************************ 00:18:48.212 00:18:48.212 real 0m39.857s 00:18:48.212 user 0m55.587s 00:18:48.212 sys 0m3.269s 00:18:48.212 16:06:46 blockdev_raid5f -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:48.212 16:06:46 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:48.212 16:06:46 -- spdk/autotest.sh@194 -- # uname -s 00:18:48.212 16:06:46 -- spdk/autotest.sh@194 -- # [[ Linux == Linux ]] 00:18:48.212 16:06:46 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:18:48.212 16:06:46 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:18:48.212 16:06:46 -- spdk/autotest.sh@207 -- # '[' 0 -eq 1 ']' 00:18:48.212 16:06:46 -- spdk/autotest.sh@256 -- # '[' 0 -eq 1 ']' 00:18:48.212 16:06:46 -- spdk/autotest.sh@260 -- # timing_exit lib 00:18:48.212 16:06:46 -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:48.212 16:06:46 -- common/autotest_common.sh@10 -- # set +x 00:18:48.212 16:06:46 -- spdk/autotest.sh@262 -- # '[' 0 -eq 1 ']' 00:18:48.212 16:06:46 -- spdk/autotest.sh@267 -- # '[' 0 -eq 1 ']' 00:18:48.212 16:06:46 -- spdk/autotest.sh@276 -- # '[' 0 -eq 1 ']' 00:18:48.212 16:06:46 -- spdk/autotest.sh@311 -- # '[' 0 -eq 1 ']' 00:18:48.212 16:06:46 -- spdk/autotest.sh@315 -- # '[' 0 -eq 1 ']' 00:18:48.212 16:06:46 -- spdk/autotest.sh@319 -- # '[' 0 -eq 1 ']' 00:18:48.212 16:06:46 -- spdk/autotest.sh@324 -- # '[' 0 -eq 1 ']' 00:18:48.212 16:06:46 -- spdk/autotest.sh@333 -- # '[' 0 -eq 1 ']' 00:18:48.212 16:06:46 -- spdk/autotest.sh@338 -- # '[' 0 -eq 1 ']' 00:18:48.212 16:06:46 -- spdk/autotest.sh@342 -- # '[' 0 -eq 1 ']' 00:18:48.212 16:06:46 -- spdk/autotest.sh@346 -- # '[' 0 -eq 1 ']' 00:18:48.212 16:06:46 -- spdk/autotest.sh@350 -- # '[' 0 -eq 1 ']' 00:18:48.212 16:06:46 -- spdk/autotest.sh@355 -- # '[' 0 -eq 1 ']' 00:18:48.212 16:06:46 -- spdk/autotest.sh@359 -- # '[' 0 -eq 1 ']' 00:18:48.212 16:06:46 -- spdk/autotest.sh@366 -- # [[ 0 -eq 1 ]] 00:18:48.212 16:06:46 -- spdk/autotest.sh@370 -- # [[ 0 -eq 1 ]] 00:18:48.212 16:06:46 -- spdk/autotest.sh@374 -- # [[ 0 -eq 1 ]] 00:18:48.212 16:06:46 -- spdk/autotest.sh@378 -- # [[ '' -eq 1 ]] 00:18:48.212 16:06:46 -- spdk/autotest.sh@385 -- # trap - SIGINT SIGTERM EXIT 00:18:48.212 16:06:46 -- spdk/autotest.sh@387 -- # timing_enter post_cleanup 00:18:48.212 16:06:46 -- common/autotest_common.sh@726 -- # xtrace_disable 00:18:48.212 16:06:46 -- common/autotest_common.sh@10 -- # set +x 00:18:48.212 16:06:46 -- spdk/autotest.sh@388 -- # autotest_cleanup 00:18:48.212 16:06:46 -- common/autotest_common.sh@1396 -- # local autotest_es=0 00:18:48.212 16:06:46 -- common/autotest_common.sh@1397 -- # xtrace_disable 00:18:48.212 16:06:46 -- common/autotest_common.sh@10 -- # set +x 00:18:49.586 INFO: APP EXITING 00:18:49.586 INFO: killing all VMs 00:18:49.586 INFO: killing vhost app 00:18:49.586 INFO: EXIT DONE 00:18:49.586 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:18:49.586 Waiting for block devices as requested 00:18:49.586 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:18:49.844 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:18:50.410 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:18:50.410 Cleaning 00:18:50.410 Removing: /var/run/dpdk/spdk0/config 00:18:50.410 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:18:50.410 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:18:50.410 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:18:50.410 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:18:50.410 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:18:50.410 Removing: /var/run/dpdk/spdk0/hugepage_info 00:18:50.410 Removing: /dev/shm/spdk_tgt_trace.pid56076 00:18:50.410 Removing: /var/run/dpdk/spdk0 00:18:50.410 Removing: /var/run/dpdk/spdk_pid55873 00:18:50.410 Removing: /var/run/dpdk/spdk_pid56076 00:18:50.410 Removing: /var/run/dpdk/spdk_pid56293 00:18:50.410 Removing: /var/run/dpdk/spdk_pid56392 00:18:50.410 Removing: /var/run/dpdk/spdk_pid56431 00:18:50.410 Removing: /var/run/dpdk/spdk_pid56554 00:18:50.410 Removing: /var/run/dpdk/spdk_pid56572 00:18:50.410 Removing: /var/run/dpdk/spdk_pid56765 00:18:50.410 Removing: /var/run/dpdk/spdk_pid56858 00:18:50.410 Removing: /var/run/dpdk/spdk_pid56954 00:18:50.410 Removing: /var/run/dpdk/spdk_pid57065 00:18:50.410 Removing: /var/run/dpdk/spdk_pid57157 00:18:50.410 Removing: /var/run/dpdk/spdk_pid57196 00:18:50.410 Removing: /var/run/dpdk/spdk_pid57233 00:18:50.410 Removing: /var/run/dpdk/spdk_pid57303 00:18:50.410 Removing: /var/run/dpdk/spdk_pid57393 00:18:50.410 Removing: /var/run/dpdk/spdk_pid57829 00:18:50.410 Removing: /var/run/dpdk/spdk_pid57893 00:18:50.410 Removing: /var/run/dpdk/spdk_pid57956 00:18:50.410 Removing: /var/run/dpdk/spdk_pid57972 00:18:50.410 Removing: /var/run/dpdk/spdk_pid58074 00:18:50.410 Removing: /var/run/dpdk/spdk_pid58090 00:18:50.410 Removing: /var/run/dpdk/spdk_pid58198 00:18:50.410 Removing: /var/run/dpdk/spdk_pid58214 00:18:50.410 Removing: /var/run/dpdk/spdk_pid58272 00:18:50.410 Removing: /var/run/dpdk/spdk_pid58290 00:18:50.410 Removing: /var/run/dpdk/spdk_pid58343 00:18:50.410 Removing: /var/run/dpdk/spdk_pid58361 00:18:50.410 Removing: /var/run/dpdk/spdk_pid58527 00:18:50.410 Removing: /var/run/dpdk/spdk_pid58563 00:18:50.410 Removing: /var/run/dpdk/spdk_pid58647 00:18:50.410 Removing: /var/run/dpdk/spdk_pid59883 00:18:50.410 Removing: /var/run/dpdk/spdk_pid60078 00:18:50.410 Removing: /var/run/dpdk/spdk_pid60213 00:18:50.410 Removing: /var/run/dpdk/spdk_pid60821 00:18:50.410 Removing: /var/run/dpdk/spdk_pid61016 00:18:50.410 Removing: /var/run/dpdk/spdk_pid61153 00:18:50.410 Removing: /var/run/dpdk/spdk_pid61753 00:18:50.410 Removing: /var/run/dpdk/spdk_pid62072 00:18:50.410 Removing: /var/run/dpdk/spdk_pid62201 00:18:50.410 Removing: /var/run/dpdk/spdk_pid63509 00:18:50.410 Removing: /var/run/dpdk/spdk_pid63751 00:18:50.410 Removing: /var/run/dpdk/spdk_pid63885 00:18:50.410 Removing: /var/run/dpdk/spdk_pid65193 00:18:50.410 Removing: /var/run/dpdk/spdk_pid65435 00:18:50.410 Removing: /var/run/dpdk/spdk_pid65564 00:18:50.410 Removing: /var/run/dpdk/spdk_pid66883 00:18:50.410 Removing: /var/run/dpdk/spdk_pid67301 00:18:50.410 Removing: /var/run/dpdk/spdk_pid67430 00:18:50.410 Removing: /var/run/dpdk/spdk_pid68838 00:18:50.410 Removing: /var/run/dpdk/spdk_pid69086 00:18:50.410 Removing: /var/run/dpdk/spdk_pid69222 00:18:50.410 Removing: /var/run/dpdk/spdk_pid70634 00:18:50.410 Removing: /var/run/dpdk/spdk_pid70882 00:18:50.410 Removing: /var/run/dpdk/spdk_pid71022 00:18:50.410 Removing: /var/run/dpdk/spdk_pid72431 00:18:50.410 Removing: /var/run/dpdk/spdk_pid72901 00:18:50.410 Removing: /var/run/dpdk/spdk_pid73041 00:18:50.410 Removing: /var/run/dpdk/spdk_pid73179 00:18:50.410 Removing: /var/run/dpdk/spdk_pid73622 00:18:50.410 Removing: /var/run/dpdk/spdk_pid74353 00:18:50.410 Removing: /var/run/dpdk/spdk_pid74737 00:18:50.411 Removing: /var/run/dpdk/spdk_pid75417 00:18:50.411 Removing: /var/run/dpdk/spdk_pid75874 00:18:50.411 Removing: /var/run/dpdk/spdk_pid76632 00:18:50.411 Removing: /var/run/dpdk/spdk_pid77026 00:18:50.411 Removing: /var/run/dpdk/spdk_pid78913 00:18:50.411 Removing: /var/run/dpdk/spdk_pid79337 00:18:50.411 Removing: /var/run/dpdk/spdk_pid79759 00:18:50.411 Removing: /var/run/dpdk/spdk_pid81760 00:18:50.411 Removing: /var/run/dpdk/spdk_pid82224 00:18:50.669 Removing: /var/run/dpdk/spdk_pid82724 00:18:50.669 Removing: /var/run/dpdk/spdk_pid83762 00:18:50.669 Removing: /var/run/dpdk/spdk_pid84068 00:18:50.669 Removing: /var/run/dpdk/spdk_pid84966 00:18:50.669 Removing: /var/run/dpdk/spdk_pid85278 00:18:50.669 Removing: /var/run/dpdk/spdk_pid86179 00:18:50.669 Removing: /var/run/dpdk/spdk_pid86491 00:18:50.669 Removing: /var/run/dpdk/spdk_pid87139 00:18:50.669 Removing: /var/run/dpdk/spdk_pid87397 00:18:50.669 Removing: /var/run/dpdk/spdk_pid87448 00:18:50.669 Removing: /var/run/dpdk/spdk_pid87480 00:18:50.669 Removing: /var/run/dpdk/spdk_pid87717 00:18:50.669 Removing: /var/run/dpdk/spdk_pid87890 00:18:50.669 Removing: /var/run/dpdk/spdk_pid87977 00:18:50.669 Removing: /var/run/dpdk/spdk_pid88063 00:18:50.669 Removing: /var/run/dpdk/spdk_pid88112 00:18:50.669 Removing: /var/run/dpdk/spdk_pid88138 00:18:50.669 Clean 00:18:50.669 16:06:48 -- common/autotest_common.sh@1453 -- # return 0 00:18:50.669 16:06:48 -- spdk/autotest.sh@389 -- # timing_exit post_cleanup 00:18:50.669 16:06:48 -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:50.669 16:06:48 -- common/autotest_common.sh@10 -- # set +x 00:18:50.669 16:06:48 -- spdk/autotest.sh@391 -- # timing_exit autotest 00:18:50.669 16:06:48 -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:50.669 16:06:48 -- common/autotest_common.sh@10 -- # set +x 00:18:50.669 16:06:48 -- spdk/autotest.sh@392 -- # chmod a+r /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:18:50.669 16:06:48 -- spdk/autotest.sh@394 -- # [[ -f /home/vagrant/spdk_repo/spdk/../output/udev.log ]] 00:18:50.669 16:06:48 -- spdk/autotest.sh@394 -- # rm -f /home/vagrant/spdk_repo/spdk/../output/udev.log 00:18:50.669 16:06:48 -- spdk/autotest.sh@396 -- # [[ y == y ]] 00:18:50.669 16:06:48 -- spdk/autotest.sh@398 -- # hostname 00:18:50.669 16:06:48 -- spdk/autotest.sh@398 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -d /home/vagrant/spdk_repo/spdk -t fedora39-cloud-1721788873-2326 -o /home/vagrant/spdk_repo/spdk/../output/cov_test.info 00:18:50.926 geninfo: WARNING: invalid characters removed from testname! 00:19:12.856 16:07:09 -- spdk/autotest.sh@399 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -a /home/vagrant/spdk_repo/spdk/../output/cov_base.info -a /home/vagrant/spdk_repo/spdk/../output/cov_test.info -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:19:14.756 16:07:12 -- spdk/autotest.sh@400 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/dpdk/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:19:17.283 16:07:15 -- spdk/autotest.sh@404 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info --ignore-errors unused,unused '/usr/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:19:19.184 16:07:17 -- spdk/autotest.sh@405 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/examples/vmd/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:19:21.710 16:07:19 -- spdk/autotest.sh@406 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:19:23.612 16:07:21 -- spdk/autotest.sh@407 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:19:26.196 16:07:23 -- spdk/autotest.sh@408 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:19:26.196 16:07:23 -- spdk/autorun.sh@1 -- $ timing_finish 00:19:26.196 16:07:23 -- common/autotest_common.sh@738 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/timing.txt ]] 00:19:26.196 16:07:23 -- common/autotest_common.sh@740 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:19:26.196 16:07:23 -- common/autotest_common.sh@741 -- $ [[ -x /usr/local/FlameGraph/flamegraph.pl ]] 00:19:26.196 16:07:23 -- common/autotest_common.sh@744 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:19:26.196 + [[ -n 4975 ]] 00:19:26.196 + sudo kill 4975 00:19:26.205 [Pipeline] } 00:19:26.224 [Pipeline] // timeout 00:19:26.231 [Pipeline] } 00:19:26.246 [Pipeline] // stage 00:19:26.252 [Pipeline] } 00:19:26.268 [Pipeline] // catchError 00:19:26.277 [Pipeline] stage 00:19:26.280 [Pipeline] { (Stop VM) 00:19:26.295 [Pipeline] sh 00:19:26.577 + vagrant halt 00:19:28.475 ==> default: Halting domain... 00:19:32.728 [Pipeline] sh 00:19:33.004 + vagrant destroy -f 00:19:35.530 ==> default: Removing domain... 00:19:35.540 [Pipeline] sh 00:19:35.816 + mv output /var/jenkins/workspace/raid-vg-autotest/output 00:19:35.823 [Pipeline] } 00:19:35.838 [Pipeline] // stage 00:19:35.843 [Pipeline] } 00:19:35.856 [Pipeline] // dir 00:19:35.861 [Pipeline] } 00:19:35.873 [Pipeline] // wrap 00:19:35.878 [Pipeline] } 00:19:35.890 [Pipeline] // catchError 00:19:35.897 [Pipeline] stage 00:19:35.899 [Pipeline] { (Epilogue) 00:19:35.910 [Pipeline] sh 00:19:36.188 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:19:41.462 [Pipeline] catchError 00:19:41.464 [Pipeline] { 00:19:41.477 [Pipeline] sh 00:19:41.758 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:19:41.758 Artifacts sizes are good 00:19:41.766 [Pipeline] } 00:19:41.780 [Pipeline] // catchError 00:19:41.791 [Pipeline] archiveArtifacts 00:19:41.799 Archiving artifacts 00:19:41.899 [Pipeline] cleanWs 00:19:41.910 [WS-CLEANUP] Deleting project workspace... 00:19:41.910 [WS-CLEANUP] Deferred wipeout is used... 00:19:41.916 [WS-CLEANUP] done 00:19:41.918 [Pipeline] } 00:19:41.935 [Pipeline] // stage 00:19:41.940 [Pipeline] } 00:19:41.962 [Pipeline] // node 00:19:41.970 [Pipeline] End of Pipeline 00:19:42.008 Finished: SUCCESS